var/home/core/zuul-output/0000755000175000017500000000000015135716075014537 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015135722527015502 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000221274315135722452020267 0ustar corecore*wikubelet.log_o[;r)Br'o b-n(!9t%Cs7}g/غIsdr.k9GfhuB?\M~:U狿h[.|yo~\n6ӟ^yzW㔮-b6"οƼ>UWm׫Y_?|uݗ[y[L-V_pY_P-bXwûxwAۋt[~ _P^~&RY,yDy~z]/:oXx$%X"LADA@@tkޕf{5Wbx=@^J})K3x~JkwI|YowS˷j̶֛]/8 N Rm(of`\r\L>{Jm 0{vR̍>dQQGFh k0&S V&@i{ C2i1Gdē _%Kٻւ(Ĩ$#TLX h~Bn%HU^L_,ƴ40tr>PYD'vt'oIEG}o٬owko%gQ(%t#NL֜ eh&Ƨ,RH 4*,!SD 1Ed_wkxdL3F;/úb}0w8_uuiBx_2dd$YLYG(#?%U? ` 17ׅwڋًM)$.j!bE"o j/oLb`r"hA ósvYļU[ Z.׿-h QZ*U1|t5wKO={mߍO1nԝG?xXW%QcJpDVHwԡ/.2h{qۀK8yUOdssdMvw`21ɻ]/ƛ"@8(PN_,_0;o_x+V| TXNrdTs>RDPhإek-*듌D[5l2_nH[׫yTNʹ<ws~^B.Ǔg'AS'E`hmsJU # DuT%ZPt_Wď츣s6ƙR*- F? xtfj>Pwȹl;͂h̞Qj_P]2@vN (C9yO|$UvވkZoIfzᑇy ^t }|#qKrdK\D2s&[#bE(mV9ىbſ~I=o嚲W9ȝQEkT/*BR =v*.h4(^&-Wg̫b]OB`i=_Z;57xh^J">CMMQQ؏*ΧL ߁NPi?$;g&u8~Y >hl%}Р`sMC77Aztԝp ,}Nptt%q6& ND lM;ָPZGa(X(2*91n@^WrN_Ŏ6W>Bߔ)bQ) <4G0 C.bqMWkB.yhi-cSDCR6"KaFٗt<>vRڡc0SAA\cH1dߛ_SRzSa™:']*}EXɧM<@:jʨΨrPE%NT&1H>g":ͨ - 3J?5OͩLH.:;ߡ֖QʡCOx]*9񣆍{v3FxlQ{jv ]t 9IhW)FRI)X)yK9ް4:ÒozsB<^+ -i̭uJ{KưЖ@+UBj -;L6l\Jk}8gf) afs'oIfz^.I< )9qf e%dhy:O40n'c}c1XҸuFiƠIkaIx( +")OtZ l^Z^CQ6tffEmDφǽ{jt'#=( 9X$=rw- >+`قSᔙD'Ad V֪v3lVx5X@O(jPcoCEeք Dj7NI0[EΰPaySwn(}+~hX(d#iI@YUXPKL:3LVY~`7KZqի8QufiŒSq3<uqMQhiae̱F+,~Mn3 09WAukYEPnO'nt3vmL=HN&DǭZrb5Iffe6Rh&C4>Qwf8*c4˥ĘP0W YW ].P!_~&^%80=1JgޛIgǽgr&P29LcIIGAɐ`P-\zʡP=_RFZ Nۈ?/&GioWiO[BdG.*)Ym<`-RAJLڈ}D1yovE^lKKiw+ڍ[I?TPht /˿e?n]FhNU˿oۂ6C9C7sn,kje*;iΓGx A7yTJ$LL-aP' sK|ۜLɽy]ʸEO<-YEqKzϢ \{>dDLF amK/0-VELJsC>?5El5uAߙXC90뼯nNNXYt\oP@gOV ]cӰJ:^q';{V=-dZB4']a.QO:#'6RE'E3 */HAYk%C6Θ%%:ò6PT:”QVay e3VWX@&avF묇cTy^}m .Ŏ7Uֻ󂊹P-\!3^.Y9[XԦo Έ')Ji.VՕH4~)(k!frH_HI\:U}UE$J @ٚeZE0(8ŋ ϓ{ %VO"d.wEр%}5zWˬQOS)ZbF p$^(2JцQImuzhpyXڈ2ͤh}/[g1ieQ*-=hiך5J))?' c9*%WyΈ W\Of[=߰+ednU$YD',jߎW&7DXǜߍG`DbE#0Y4&|޻xѷ\;_Z^sнM\&+1gWo'Y;l>V ̍"ޛ4tO,{=hFѓ$b =D(zn;Y<1x~SJ^{vn 9 j1шk'L"cE=K]A(oQ۲6+ktwLzG,87^ 9H\yqū1)\(v8pHA"ΈGVp"c ?Z)hm.2;sl$瓴ӘIe~H|.Y#C^SJĽHǀeTwvy"v܅ ]?22R.lQPa ˆSܫ1z.x62%z].`Gn&*7bd+, Z`ͲH-nမ^WbPFtOfD]c9\w+ea~~{;Vm >|WAޭi`HbIãE{%&4]Iw Wjoru ݜmKnZ<X; ۢ( nx K8.|DXb +*598;w)zp:̊~;͞)6vnM!N5Cu!8Wq/`FUwWAֻ,Qu W@ Fi:K [Av*_958]a:pmQ&'ᚡmi@ zF(n&P;)_]µ!doR0`pl`~9Fk[ٺ+4Hhao-jϸ??R<lb#P-^39T|L /~p│x@Bq"M/lja\b݋af LnU*P(8W[U6WX ZoѶ^SH:K:%Qvl\b FqQI.ȨHWo;Nw$͹O$oEE-eq=.*Dp,V;(bgJ!gF)892sw*+{[or@x,))[o新#.͞.;=fc<)((b۲Eumw峛M2,V[cm,S~ AF~.2v?JNt=O7^r.@DEuU1}g$>8ac#sĢB\PIPfwJQJ;Qxm &GBf\ZA$Ba-z|A-I @x70 晪MV)m8[6-Te@`E|=U D(C{oVa*H7MQK"<O%MTTtx袥:2JޚݶKd7UZihRk71VDqiގ\<:Ѓ3"gJJčE&>&EI|I˿j2ǯɘCGOa9C1L ={fm&'^tigk$DA' elW@Tiv{ !]oBLKJO*t*\n-iȚ4`{x_z;j3Xh ׄ?xt.o:`x^d~0u$ v48 0_ | E"Hd"H`A0&dY3 ً[fctWF_hdxMUY.b=eaI3Z=᢬-'~DWc;j FRrI5%N/K;Dk rCbm7чsSW_8g{RY.~XfEߪg:smBi1 YBX4),[c^54Sg(s$sN' 88`wC3TE+A\.ԍל9 y{͝BxG&JS meT;{З>'[LR"w F05N<&AJ3DA0ʄ4(zTUWDdE3̻l^-Xw3Fɀ{B-~.h+U8 i1b8wؖ#~zQ`/L 9#Pu/<4A L<KL U(Ee'sCcq !Ȥ4΍ +aM(VldX ][T !Ȱ|HN~6y,⒊)$e{)SR#kהyϛ7^i58f4PmB8 Y{qeφvk73:1@ƛ.{f8IGv*1藺yx27M=>+VnG;\<x7v21՚H :[Γd!E'a4n?k[A׈(sob 41Y9(^SE@7`KIK`kx& V`X0,%pe_ן >hd xе"Q4SUwy x<'o_~#6$g!D$c=5ۄX[ു RzG:柺[ӏ[3frl ô ހ^2TӘUAT!94[[m۾\T)W> lv+ H\FpG)ۏjk_c51̃^cn ba-X/#=Im41NLu\9ETp^poAOOgJ8@o2k'Hr~4Z(I8!H G8HNW%1Tќ^?xBVQXodՔz q[*ڔC"1Ȋ-R0ڱ}oF4 3vFf#8^Vє+k@ :)@%9@nA B q 62!/ 6G (" u:)fSGAV(e֖t܁ ft~c.!R0N<R{mtdFdHÃФsxBl] " Δ<=9i/ d ␙F9Ґ)Hnxps2wApP!se]I)^ k?'k:%Ѹ)?wɧ6a{r7%]_Ϧi~ԞnZhubW*IakVC-(>Z#"U4Xk1G;7#m eji'ĒGIqB//(O &1I;svHd=mJW~ړUCOīpAiB^MP=MQ`=JB!"]b6Ƞi]ItЀ'Vf:yo=K˞r:( n72-˒#K9T\aVܩO "^OF1%e"xm뻱~0GBeFO0ޑ]w(zM6j\v00ׅYɓHڦd%NzT@gID!EL2$%Ӧ{(gL pWkn\SDKIIKWi^9)N?[tLjV}}O͌:&c!JC{J` nKlȉW$)YLE%I:/8)*H|]}\E$V*#(G;3U-;q7KǰfξC?ke`~UK mtIC8^P߼fub8P銗KDi'U6K×5 .]H<$ ^D'!" b1D8,?tT q lKxDȜOY2S3ҁ%mo(YT\3}sѦoY=-- /IDd6Gs =[F۴'c,QAIٰ9JXOz);B= @%AIt0v[Ƿ&FJE͙A~IQ%iShnMІt.޿>q=$ts,cJZڗOx2c6 .1zҪR "^Q[ TF )㢥M-GicQ\BL(hO7zNa>>'(Kgc{>/MoD8q̒vv73'9pM&jV3=ɹvYƛ{3iψI4Kp5 d2oOgd||K>R1Qzi#f>夑3KմԔ萴%|xyr>ķx>{E>Z4Ӥ͋#+hI{hNZt 9`b˝`yB,Ȍ=6Z" 8L O)&On?7\7ix@ D_P"~GijbɠM&HtpR:4Si גt&ngb9%islԃ)Hc`ebw|Ī Zg_0FRYeO:F)O>UD;;MY,2ڨi"R"*R2s@AK/u5,b#u>cY^*xkJ7C~pۊ ~;ɰ@ՙ.rT?m0:;}d8ۈ ݨW>.[Vhi̒;̥_9$W!p.zu~9x۾vC;kN?WƟ+fx3SuKQqxST Ζ2%?T74a{N8;lr`$pZds=3jwlL Eڲ t|*n8[#yN SrA GYb8ZIaʼn8 #fg3i`F#5N 3q_M]j 8E!@1vցP7!|+R@;HspSI]ڻCZUcg5pDcIϹ,oN-_XI,3\j ]ٟ5~' SuipA!C厐$&k7dmhz/#"݃,YqCL$ڲ`"MUbeT>Xuv~4Le͢ }UVM)[A`b}mcE]LCEg=2ȴcmZ?E*-8nhױ1xR2ϫCya` A y!?h!9yL%VLU2gr26A!4vbSG ]ꧧWp/ &ee *w$-`J\ ptǣC^p#_`{ К8EW>*(D{ٛ,[fnY𱹞M=6&$<,"lX-Ǐ_whaE 98 (oѢ/Р΅ 7ցl6618ł_1/=fu).s¯?.S[{'g=Ҥ):d8h\y6]t1T7IUV:;.1& ,5΀j:<< +Y?58In'bXIǣO{&V\DŽ0,9f O_"[l:h¢8wݓ19\:f6:+ .3}=uvKc ٹeS<>ij(o'ciS<{1$E[nP b?8E'xv[K+E{,Qƙ1*dcs_Z'407|qBOgYU|U--sG8`u! qGYܷw;ȌCPc_|(RaIBKb+{P.T! =ĦiTob d<>SHr][KqWs7ѝBYǭ~RR"p9dFg|K- obY_vM 4>/]e/dy,8!xŋ5 R<^mYo 3c9(F?he:9[_v~\:P ؇'k01Q1jlX)/ΏL+NhBUx~Ga>Z"Q_wjTLRˀtL L+BT҂ll魳cf[L̎`;rK+S- (J[(6 b F? ZvƂcW+dˍ-m𢛲@ms~}3ɱ© R$ T5%:zZ甎܋)`ŰJ38!;NfHohVbK :S50exU}W`upHЍE_fNTU*q%bq@/5q0);F74~'*z[\M-~#aSmMÉB2Nnʇ)bAg`u2t"8U [tJYSk, "vu\h1Yhl~[mhm+F(g 6+YtHgd/}7m]Q!Mę5bR!JbV>&w6οH+NL$]p>8UU>Ѫg39Yg>OF9V?SAT~:gGt $*}aQ.Zi~%K\rfm$%ɪq(%W>*Hg>KStE)KS1z2"h%^NEN?  hxnd/)O{,:خcX1nIaJ/t4J\bƀWc-d4M^d/ ʂK0`v%"s#PCoT/*,:[4b=]N&, ,B82^WK9EHLPm))2.9ȱ  QAcBC-|$M\^B!`}M^t+C~Lb }D>{N{Vt)tpDN,FCz~$)*417l;V iэ(_,j]$9O+/Sh]ice wy\Mڗ$,DJ|lj*à␻,?XAe0bX@ h0[}BU0v']#Vo !ې: Z%ƶ(fl>'"Bg< 0^_d0Y@2!ӸfZ{Ibi/^cygwדzY'Ź$:fr;)ٔf ՠ3Kcxwg*EQU{$Sڸ3x~ 5clgSAW"X Pҿ.ظwyV}̒KX9U1>V..W%GX +Uvzg=npu{do#Vb4ra\sNC/T"*!k愨}plm@+@gSUX覽t01:)6kSL9Ug6rEr(3{ xRP8_S( $?uk| ]bP\vۗ晋cgLz2r~MMp!~~h?ljUc>rw}xxݸǻ*Wu{}M?\GSߋ2ꮺ5w"7U0)lۨB0ח*zW߬V}Z۫ܨJ<]B=\>V7¯8nq~q?A-?T_qOq?5-3 |q|w.dަ'/Y?> (<2y. ">8YAC| w&5fɹ(ȊVã50z)la.~LlQx[b&Pĥx BjIKn"@+z'}ũrDks^F\`%Di5~cZ*sXLqQ$q6v+jRcepO}[ s\VF5vROq%mX-RÈlб 6jf/AfN vRPػ.6<'"6dv .z{I>|&ׇ4Ăw4 [P{]"}r1殲)ߚA 2J1SGpw>ٕQѱ vb;pV ^WO+į1tq61W vzZ U'=҅}rZ:T#\_:ď);KX!LHuQ (6c94Ce|u$4a?"1] `Wa+m𢛲`Rs _I@U8jxɕͽf3[Pg%,IR Ř`QbmүcH&CLlvLҼé1ivGgJ+u7Τ!ljK1SpHR>:YF2cU(77eGG\ m#Tvmە8[,)4\\=V~?C~>_) cxF;;Ds'n [&8NJP5H2Զj{RC>he:ա+e/.I0\lWoӊĭYcxN^SPiMrFI_"*l§,̀+ å} .[c&SX( ( =X?D5ۙ@m cEpR?H0F>v6A*:W?*nzfw*B#d[se$U>tLNÔ+XX߇`cu0:U[tp^}{>H4z 4 (DtH-ʐ?sk7iIbΏ%T}v}e{aBs˞L=ilNeb]nltwfCEI"*S k`u ygz[~S [j3+sE.,uDΡ1R:Vݐ/CBc˾] shGՙf 2+);W{@dlG)%عF&4D&u.Im9c$A$Dfj-ء^6&#OȯTgرBӆI t[ 5)l>MR2ǂv JpU1cJpրj&*ߗEЍ0U#X) bpNVYSD1౱UR}UR,:lơ2<8"˓MlA2 KvP8 I7D Oj>;V|a|`U>D*KS;|:xI/ió21׭ȦS!e^t+28b$d:z4 .}gRcƈ^ʮC^0l[hl"য*6 ny!HQ=GOf"8vAq&*țTOWse~ (5TX%/8vS:w}[ą qf2Lυi lm/+QD4t.P*2V J`\g2%tJ4vX[7g"z{1|\*& >Vv:V^S7{{u%[^g=pn]Y#&ߓTί_z7e&ӃCx;xLh+NOEp";SB/eWٹ`64F 2AhF{Ɩ;>87DǍ-~e;\26Lة:*mUAN=VޮL> jwB}ѹ .MVfz0Ïd0l?7- }|>TT%9d-9UK=&l&~g&i"L{vrQۻou}q}hn+.{pWEqws]]|/ǫ\}/J.MLmc ԗWrU}/Ǜ+sYn[ﯾeywyY]]¨Kpx c./mo;ߟRy*4݀wm&8֨Or4 &+Bs=8'kP 3 |}44S8UXi;f;VE7e4AdX-fS烠1Uܦ$lznlq"җ^s RTn|RKm;ԻZ3)`S!9| ?}m*2@"G{yZ${˪A6yq>Elq*E< NX9@: Ih~|Y4sopp|v1f2춓tcZF_q|l/rY ':omQ#Axdc0[NJ6Ou4=>DUX(;JBUTFY:^Dƪ0c)RYUId{Ǫ9 KiXu U! R"/R=66BcURB..8[z9J 32,oxJV>))[arx,񥩞hZ` V^j`XBY\Yp^Wp_7|^/+҄*PC]g6msq۰! && t}7EUYQ\B[*ptgeY;W U\lbt_ZrD^`4fY%C|j*Pg&P5R]O1@W\ ط0l:"0',]/RJC!Crn3_tuywaXSC\쵭ͣ?: jԂ«G,(y9`+Suñ-B%Hrm=V_>``8~d붡Xo|'*t_^ o1آ@(!ˏ ƢT|޳ZGJD2 ʨU _L,|>g 5߹ ^OWKE^*_L MDZY躮a S 4w 7/Tu0-kzf1pB@\cq8!a?LGE []qxGi2- `<"K l= B܈!Xfe1Oe͚RFPD,E^|! 90*_MI br1OLPWT 1 (Iy|eo_eK `N΅,*/_.ςrV5k[bu^EHu6 bZLZ`xS˴tVR'RXu%g)'a,X'f[eV_0>1 Ivb%djQNGiYZu\](aspP0RL''Y+Kr?6IFD |Ƽ'㺺aN?i9γ׺=9xJ %C 2K "zdA#:&+VrFOhUZs=wm^s x1G C{RiOL[e= ϒf\P83,-X5}:\uU3.iȜFPIY1 Bqq՜}}>:+.< C(W+?5AUE|ۨd~,:7Ҩq!8@X|ЇUQ#r$FuyOg&ܝqt <<Z-1HWw%+qįRI(iP\fI\UyyqixɅ4*oXp YH8݅fKp465iu~}bU3OP-i`v7'o 0dtiه㳏XK+o7-KTpx?M*`!(=="lrl+-n+@- Tr-TTEĚ HSwO@_S$MU)K!ǐewy9vhwgQ 7h S}ߜ~*s(swO{Vk[g!xE^k- k4"HHP'ݚ]+Ǚc϶9iWQj6 /,A+JK CyrU$(F8_ɻβ*S~EڤC$H篟>jܒP9:|/i@,J<#7TJՄrr @>-4TBg>/o 5M7.W rᒖ˔%QІ0Vt?B z}N; H( xxNECP4)Ik\Q(M%z6=8v.E\ e!KLwwJ3ZҐX[ OG5[65!~{%QXdQ8h> Ǖ(,,/`VavY!NS'BqP ~=#Dl2ݾGC#ͭ<*wO[]F$ڦ* oÃeX[-`h4sRES8 }Ө24Jidwuj25GFңs,ۓ2Ge͒',l)ODn:4J!=:Ư2+nnhnrrm4YD^_%x_OF޵|gA4_= 4@% ol|9mׄ+)KDV'H-ՒSƈ"9j!BuKVT =#2D -̰LCN8i 5HZZ_J-ֲ)7`6N-YgbdyU$ [A¯sPʂ=E9qo+xϚiZ[m?OO "J uS>F+hʨ_apE9@aŖe*-o9FDR )<X!H_7o_ .Dciޖ0'$jZR{(rm[ͱ\B Qwreʶ-aTe bFYnJ`"/7!܂-F7crihfXKe0R #ya Yw5+mV=w 5-;kLKHKEwTmxizEVW$3"$_]VF8@tp^; ,>Y܎ V= aWEf4 򵫺6@?,3<Fڳ{.A,zwsHE.G!;ֶBf 9my!&\]fȎ4S#Std)^qMCv`0!y}jJ {hXi Ӛ4l2z |Ӑ}ʛYb"F)먉hOϚmI¾RTMN.o/R.rZ?Ugq6_DiNy4dɾ#_9Z%%8DTeVZKw& 53ewJ}\Y6ndsoid)\ 36>~wik3Դ 5żOh<mf%\4&&< Ӷ~M0 ~}+A}M缍^X'yP[?8S{e'{ozBA{B4* S罸X^ ~N l~@͍U꜂`2(OdP=jU~0'cPP0 (W'oi0r$Bd~0sqJO(~,CA̔X}Y=үOըۈSX&qEж }ήET1Fd>y"h} eP}͖ڳ;'g ~ <υ ec%Nx ewO9 $fEݜ|`(5SW8 Z|~{.+G\fTB OX;9+gרj<N][K%KnXǂFޣ*;FkQr! 4{NԻ aq,t?( @Kė#nٻ޶q-Ww `67if,ɉ+IACR2;m9{ %!yHФݤm˿js3D dqjR)Rҿ%ɳ )ԎxvMfBG8hfEpH[jIGjI@[dfd3KeX7k`CJ*sDsrO&aZaѿ (U*sZݍǻ$7IFokerM)?atBFur.75ey5^r9QYpJG`w%$Cy7i!/72I135Yqbځ#7pԙ'ae_Ϟ.\ˏF@bxi4!ף&yVq*j 8iΣ6ed_{l\m0- G.Zuc̖K uVD`L[J i g'7H1^)6|Ge)z7-v$Jx@R\0T!yD?0r&HW S)w ^+\^.7t8 s.wbXuYŃI8!W otmKi4eay+}w`v@6G؝uM+vqSM+-88x}\ _ydf a)NdXD+0ŚpǨa6SYU=ɗ)_] _vyK/KX`t- A} {Uʟ\[ .<2Syz͆|i, eN`H7d01]"Gg'Y' 澄m qMfzyh4=*I#8_؜2&Cq}z{G7ȻOOa<{8LS}o; a':9C[~~u]<80_UXĨgXE=Ѝ\A"0BnǦt}?9z!]7|a^ 0óԵe x0 P akR65A'KTN,5(H1tL"} $CSӣI~ Ku)48]-Q4RzK(<ԂqLLk($& ZԷ%ʚ7f9=8>E|2\,FQds 4(S| j>!>MR2l ߫r$ 6AgXaq6E70u>큕nVDDȚ(8Tec(0.XڥKBQwuQx^`~hƲV8k0>*}g4]Y{pofۦks7t- ƽno.ںBf+&} S}ߖ]ʇIA_+B Or8Jh;Uoa]j%hȇ8Մs%хMnQ@Ϡi#GmƙP["HiINI9.BXᯪPH(]P>t;B醄u}ԥt>t*ݲWJVnҵƝNEjIJieۑ6AHZjmH|P|}Bv kj/jA>vjAH:lH"nG!zzkOކzk/A>v tN86N1(LnT P,ˈpXTswxfl󴗨+]U_y2 t&n2/t:@rNoٔD9pO!ta!InII!] U;`RJ' gT}gv9s{cd_Ϲr8&昆is_)yLL4}GSML̨51 WBѥ pW*w .\ͦHX(ZrQ)P ҟyLJ@}>N?/$ϳ"EhVOU]Poa(E:hqr|)\"$Wi`hH|++9\:|Eav|^4d$DS2/HHI}`<8I՗^9JAJ%L嚶(㿌~g# |^UR/ -.")iKn=Ncx''q2 g#cVpe,["%cʆy%%3&y,جdqsX ̠-fch/$G]~Q,|0Dz46)7aWG3A?f DsdC<&4.9`v4;OQ%_;r;i R}]$}Lfç^P&z!StRy/J93B; 5OcW?_Td)#H$'h:US(Ҹ|SlZ\4qLpȊqǬZy, D:DGF^f_o8St:S3*|AML髮\ &O@UH1 |Qf&|@xޔM) F( RmbWhōqx/GFI26>.O}PC:Dr; _d< y8.Ox #PmI@Ѯ!?ZxfoI-W?_;'-!3<& ΢4ܸu_3l e|q+t^,p7,b% ::/Ԛ΅%L?I@=WhwFޚ;'+̦.c;oTrwy1GAqo}GM He;a{n<63s pvNփwq=qNO{3_ޝ췦oos~)q<ӏMsg%R -J{EZyo뙚~/w_݇  xwWw3J>K!|n̷ٮ2 "󻸻${Б.D uOGSSN{51Z`pϼu=E`j ܍~=?HbGf94^4qN7l(>(d$jMGy80X|:|foY>`P MƅeK*A؏0|-~ˆa!^%!Ե2rds\%JZIٕ^x84Gj7zi١H0c7k}Y!3,Y:岍*'$vl[CXl.jHnK[״5#~Y|iyIc]P[6NAKw1E_H!u5b?2#cDt%g=A羶^Tpb$CoثLZLSC1^tb ˒\ʺnHG#C3U8I4KkOA7|kk#,zq\X(AW!šrHkt# v>65kX瘉E I7$8r/Z7Lkd,cހ_'$G+wQ(^չ>ԹOes`IYcBY3~GRYiU3) pcxj;V"oY#E/I- v,DEeg0\Oc$[=1.i,Llaʀ[&7[M*,i1p|>]Q }s3)[~Jeɀ(k' G+ܾIpTc :S oߪb2XZx0XfGٱ >UQru 8.=&(ESTZ!Ώ5\ĤJ1̋RV6d2!,#cZ4Wi , bl m右ٳ0!=LPjrl(,^Hс se<>cBvWMS9Z5>e.%Bc"%&t1j7ؑNV \nJTV7W1uD`#ޛ7) EEq{M=Hs + (_5I1[.8ƹ=X7 ' B0bhƙU, _oޛBYx"۾^/5[>#iVw'+J*^;c َ'IucCިY{ SŠ1(Qƈvm PF8`L(!N噅reB6mCHZQ f,C*ܧkd ,$(yYyHٜ'QPG9եIB&oݯ!vP)h aJ1쥦TYK,nICk,~&hJx18񾲻5Ǡp䯧5n^1|tPFCX3;SFSqUOcleqb>=kH&6 \fΣ ITQRI[Mr,wzw沕Yiý Xհi=hs="ļ_Gr} #~v0p|Qa_Ap1RҬ따toMy#<8hp\4Yt$sPM2C{$8gjc|bRVpJ-DT*F ufFbKGP֛hB_v z)FJSxլ` /B5 6#6tcs.et,'V(i-0Xt>lYn^: ~N@j"&Fxy `,F{x2z6Ffrq.R3KQeFDݤ5vhͶSViWoH4 s\DT޾6Ip\?34A1X8b0>KmV6`SGŪ“Sup4yҞ߳~/u}qtض`$^H,@L_hG8Xfc2nz&bhÊ ס'נLEzIn4J*fu.Oqj28xvd{}Jb|ǘo{)vC-ݾ*@RY:ˤ``#Ol$x&yF<z"lkfHl61M.o3 j\jd-52 4"i4lcİi= r+ۆW|+I@7"yEL+*q5a=%o1%1dFX0jIq, /E[T eX 7F.]ACi~s~-q0bj}_[@6Vc$Mݭisw}w73,<{/̫ZU*yR#l&]UNA%%YvnRȞ^t3v{Ip:xgP7Zj$(W3c.@#XX[D~G.vy`փ롋,ۘ0U[DX<GOC*|=ۀ?STj4R|Qz̏$8N/x0q,=ڶho$8^,DkQEfT6y0 Ucˉƈ!@QnH6LjicTݮ8xe;a[$y8x9.}usԜQW^½~U|*n@}2/)Ayɞ-j].c"f?᰺ PH 94g]3BAl=r!iȥ880xP3-2`ZixaWH0 ެ3N[+n V~d.mSovD gXBlL^Gp>݄!Wc$m≬w G4WbKqWcUƼ&'n@8'j|pR<Brmnz?$/vFh1؛Y: ?['N~xjڦ~!ժتOu$U> BXSbMV4,8!G RGcIW^-  R; u+Ū OCV:PN%@>g%;JepͪҬ2'2 Ac&VLĤɥA9zIpo:ycsf8P-9yrĨy-?0ޞnAw4}~?'Ѵs662U JH:VKoQ"  %-8"4 fVxpK$"~l7Sb:bt&K'9X_*mIלB2V]6evΰ"DhH!Cl)e#ֵ |Q r󑽎R'qp qmQT&kWQtu67l#M|x\i5qwWj/ʁ$^1~]ˆgCew90&k fƙO\Cϱ®{h>cS@Ī\홛q/0|k_ C奉Y&1{:ؔz؀HP4/e:k#u$xh.bR~lOY,J1!e9Q* #mӃyy8np74gH#nwV ~4\ T iׯ/$8I%=e_t9u6y$ Kg*c<S,e#9_~(6o^(v }!laDu/Ū|gFG]E"/ב]_$r*vp~GK}cG#ij?ti7qו'.:IUP-#SzpWF'}S@:('xa#-o$M}b F׵X?.Fd[) |||/_բVee½|yʲӃ_-PH8 7ʼ(}YA\\/y?gG}r \ ?EJ^R)0]7GyVN#4śTU^ܰ7X.a{0Ũ|5yZluFќdñy4cd4{zp~աm"Q"%f4g&UIB7GU^Ͻ!Ec5NT^N_ ; lgEl8=׆F~SCwo0#㲞^9z J~]OnDn|]>3#ܝ[<148 ~c:~Xe5\,K}a 0|PE& >N>ة'w^ "P'J (SqOm/&;O@x<;''(=ڳ#B~-q'|Rh};.)p&#~StЉ'#1"NJCZ o?x z.H W3W@\ڿ2~$:RJ~6c{u+_AќN*z0c?# , ]EKGv2y,h Ⱥ{/A)l5%Fw,Uޫb싰kV_\/荜I"@FPeAQETa? :67>vha|ǭȕ{sFeIյW@+#<1gדW}گ;(SJn-%ErfܮՂ{ ]}pG=w63ter7Rd#܋򦰨qy b/"qy6reK0t }_]j?|DQwQ*ym1] LgJ??/UAz؉6v5yHڀFdիw@_-R!tsۤ^NNf<dQ=|Ҕ@ 0:s7ǟpaEw{+}AIQW>^ ӂ6e΀`Dk2g\ڒ9j-ZI*U~dp$QH-E9xπb-tBGo\x )&Luu观fUA#Z4º=AW8*s)Z$@hArX=~ Sq+3 h .rBpFnbig:&;p@bh+*HݞsRzyd ^Su}Z8dM7O`m S X1x&ZpInMk RHiJcml4h>b췓/n·Qu] f= &F vICHӭ $ZEC5DUTQ;f2'/dF4%Jjl`U=&E@&S=kcQuPjJFUGUm[% HBg :*1kL*ՆmZj*!uK;ς ٟN|uon/o%EP)ð_~q_ ~IXIځ+kI 9')TXc4AE'H)ʭhA MVb_~f@qfYPi<5zbDٔbKʼnズFUV,i~Vlmj"8@nt o#&T4a =mIߚpiYs';Rn;WLc,PʓH@H$ J>693t+YbO1ɌY;BOIl7O-wk9&;Kˢlck3I bb BE&M~ڀ㥺ꡀuYA#b}?:{gyA.{:uY YY}ˆs_{#wsr6. 3g[lV31Cه텳$+A$hnnH[]rBnݠ81i,RȜtc8YP9 cgbMߘxddsIu]3"BnxYm­0&D5~ADjH X% SV8[<)jߑDR6z>R>V3TǪ)znHyWUa3:YJƴQeTO>h3}$1GtƤ۟[MK~|à>w7!m \QR9^^X>]ߵ?#-5gZPXu1BDa}ĊG< tCha@ǟ(]1\dп_\]E8y7ϣx2^β4H7a)1F9'1qƘh("fQ[.SY?=nG"?ޛ9s2{xk,N4$׳v5$XVV mpY"b,mh8Xԩ4eF!BJ&feɨQ-ɨy `H2sFgh3JJK0Fp 1BTPT-rxuɨQ=Q!4i5YVT F [Fc@ @BCj'1O5 s&_Y2vTK2v^C2 %Uɰ2AIJe,S i-qQI­,u;%u; ꛳]mľ5إ>a=rՎ7b_Z%b˓]ȳ:c(Byv9N>faD)V4A qͣW' ݹh8DґǃHf%Bf&77ST]ɓ81bb[&KQ;`^NܟìH\GYD"GApA4_yynyLT>Ï0>B*B `}LLH(\U+A.Zd׹kSv888 =`Ij}2B2 {YCTnGЏi;Wsk}/<]7eSw %w zPp}J1[xSmB_OuaN*z/TSJ]`}:)-KҍQr6tp]1+ W^pnzT.yp:tYt uZRL\|KU(ׂ!iVTrY6~]3] HN\{ieUЬ]9 T.Лy'Ү2S;:{l_`0o}4׉?Wue|x Fؑ{5sW X2 ?6Ml2^v9G$H}aTYEp Z?~x;W~5yUXߎk9޾w%?6 SJk @fɟG.CdRcλ X*ĥfj*r^0~vc JA-$s:c"> c'$g)ϛ_O.)޺ԍ\ߺ/Iܼ|'|m~qq^Mu 3~&/C8\L~qӔ󗗃\Su``+t3ǿhcvVP Bӿch_ 4Cwa&~jtychMכ;|#5{޺ 5ݕk} W 3?a [faIT~",l/n'z+ 5> ,W:]>M;}ߞ )hHLV |%+M8!0cFAs;냛zd.nئ WLmC]%hk8qe=r\Z>oV)i&(&41eqT (c2qilqĤL-LԆL4qViF (mnJ6$ݦyNbV':#SCSuȤd 5~˂ܵxׇ]G. rסwQh59@ A"gj{8_!l`Lg{e8-yxDC9[=CJR3c8OWWW?]]/"Q&M3LRc6EREk QKmCT8mZɶI?2!:12"YZLr''Xڀ5*Os875Dp2\$Y Xf":a; $jLЋHėHSh[Dk1qě y*}x*VƬz}1qql,c_nLfxK"Q1ĥt:vdd23,3rDO0e[qP30>6t $ tg3 2{ I]FIm+PJ̲D$v(+%VIgCڙK&rҼLA0Ek.k-E*T&Q$2.R߯!rsC!SO&$ 83"%2O)hR4Z] ,E\ڈ8gd|4qFku4YSQNr7ёK%P|[|#9)TܡMaa&B)ELY,I9%k'L9U 2)/YƀGt|\:),`haԺLei)*H:eFY&4WlȼhR1.&K30%VxRd=Ŝf}X80bQjQj#6G(p(%{ }Ä6 HãF"!0WE^OKr<:Px͡},U Q L2>ljit??n*t%?}JkxWLTtV-Z*|CO;?́a4Rj"HFe"6i,f c +eɇqaBGTV$&8>>EiiYEg3eGY'dDD&I2LS2%5;DNHI~bDXcH|-,Hn᎕[-gB]MY^_nrmyabaZ0]a 0B۞0-冡U`$\y+RtArT}wڶiHi 7J"ҢQxW<$>H+jN xG\vX.kh8]lQ=/Dsй7,V9DDi&*ReaPw9xp@6Ub,DeYj jFE(Li̸"39 4#Ɲ4*U2%!Xs/Έ")c-c9@LTH4&b : kۥ1Ahr8_?\Ec?ٜP]EKЋZdCw\.ac&6m{9+ e#hF ɎK쁙7ڶA-Gf2P1V'vn |/nm5Q :YݔJQ >f4J;+91Dg=w/]4_adZ8>3EȰ4!--7۶c=-Aa>hVB yh @J@M0E`y*ȘFaxzy%ı%Lcrkm"ckZ=<4(ӣ}ֺg HxZz%ߖ>/h5:jXͤ&C@}!Y V7VÕGnoB|mX8d݋w!v hH'v:I돻h1RY VhF޶G. hȚhWE!Ʀ>~26Ҍ{`1@FF c"43eV.V !߯2X ̭E)f1_Y}QϴԫWMML=9ǶB`Y`-e +Z=h՘U f ɉu0$C9[idd,wVtm^L}+X=apRL&ZMpaiZ kI4G; n@[w!0{6m{Z֭'B`ZmQZBJ[O3p ֊yҶXmVY%K7.B%"0U^!Jfz q/f}(Bh OB.t׺ViޣN5ެu nڦǩE6 ;>Fw??蜅:o}M,=>XW\CfRa nԀ)3SJ[$fwVƴU+&Staeaapbf/Yk6@F&E2i12Ubg4 ,dž܏Tt/ ȎB*.0L+0nOUJ?9ePҿZDў_I2gb>Kn*6'rcKxK}/^]]\1s} !$r4utmÄjJWo7,O8nݾ*uuq"ŋ gpM>dkXGQFq}8:T&)0ߗ TU?@Mu"bAg4֬X[ǁoG`Vb;nEҦL^etgubQ椕u/bYog\^ДL?ySŪa |c*)ӏsrA48V?w泋D2]s=cQKTS 3>$E.|OBeO -hZ[WyaAzjE3/F6cp#GKo՝Y3Ty:Ig:-҇J_~ûa%ΊT!n^*+%z{lziÆ;uтjL-@TX{F\S;aftz)YòSZC6E{O I[aÉWU&6~!ǒ?=for#AW@x}-yAכ ]!!YJljB?:5|?Z|RsԩdФn0Ⓐp"(ՆloBcVzտj-cO ִmHhy%жNbnzO9]\[IT68a`t>̫bَ7wNhŬn2b{@вaMVjL%,[e=O<4gt9InI;*/u_d+χ?K];VD*y07#Q3pt+ǍIߞ]UrJ>tTn"53j5}ߛ>¦˛|n՞1/c -hzHKL+bhc& SU3"4cyjyw7hqqa5rY4k{OF._YV24}LWB6Th;k>X˦hX>wG}36wTz Wn/x3,?rtb]N 4 mrqysiR"7_/]YKWE]9%?G9_ᬩ$WB3VU tvocՂwvĚըBqжvYvZH5Ah}}wk} @vx*. )ؖJ צ [n׶|րZ^Ǹ//%i P `e{<ܖck޵=[!|SfFdc`vt>h"aU[}߂`Z![/y˦&/#s%'`;g)1(tCP*3n4 Qy-x hUXU• Oڹro!f*3u6V4ť[;$f$Y/8(祩R_;)-s'cwaO:_V1&lmqu],UJO6m-"a'YAG׫ED4f?bҥ? @6;$d\qmӾu\\~BU`)@jY rW-Ww}V`U!%*Z>JՃp$NelCx:X(;2_G~<ݿ^t{N Wn+Zj_Alگ5dLnWT&jqĔYc-V^w"̄"wJoG*"`'KZ?3alj"lVvˣރVt)yxĝk/{J Z/ش\y,mV k;^h8`68|z2htw̒B)+_n IBy\dELDKc83;rkm"ck;l zR6#wLZqF-MCf/=Vzf$7wE^s9HvZt(%Y4Oh uk^,pi&I2L! S2%3;y" #e!`!Ř5B-5p_}NSzч"@{ z C[oAJ x[ԶK_nYȖdf]]oN`HO>XNJk(? 俢߳lJ"?H|)t#ɓgjю]LW*MYʏch| EZ)43`\rث)|SM t[dڧdamJWLBjV)YV%jU9}dݥdUQr*5}fEe*{ //)r/~IE$ɍ_thrx MhM!;Ym]gT[^ jfcrE#6b;~2{J-tHYcRZt-B%QԄgRqz;sTttx!ο{Gܲ8s k\4e~*#.g)lu3yїo7^],'g(""R?/1љHʳUveYMqk̽_a yfvA9*#17I4Z]Ϗb!B*{Ҩa022Fvt~Vg~L:7{Wȍd`5`V>X.Ƴ01Ӟvl<˭:o0JYM)lopNEx$_Aj6_nfY ts9Z-R`)~_` i+zu;4|7,,P}*'仸a/,g`i~t 4]6tiڣCְ+ut,ןŨtVI?8W:d+n(}d8$x^Ce7_U"[fdHbA?}^XvL@6}>یMxv3gPn LI0^w%) dkM ʭ襼w?L}4V)jX5jʜ$cqk$26$D16O9WTrD̸"ABk-g'5O%&3ƭTD0HxBcpBAԩ{o;y0ej-*g{[K(C.&97 c}(b(O˷v|8Tˍs<Xl޶?/hyޡCx,Uc%g_[4/ٲ.*9'4:y/آ7~U'5=̬]MOڻчçWGTmh[s\)u?Jeg:C^/YKsYf1D ҉W;H16ZMa|j_ ޫ%&ُL.wq̨N(ogލz_vfpimhBNK!?ȟk>PmV bo0;oZcݺx6MIQ v)ztKh?J[-O?5Y?oC?NTs= nC.*rk俛Je\y!.mͦF}l*a*nrS /fP,]"U]?>oJa=MrU)w TgO/v^j~)|U+Op\;w|hQk/rUsx>GܔєW7?5)U[Sok!O"VPo=zo0iH,_[d DkՎDx8Dnc--c,5ۦ9yr7-f29\6 WrŮg'8suÃ!}Mt_rm򃜂)~{QS/;#%eƏ,3M8^rxI9A2Z>u.ͱoQ=cD  3RϦZV?_??.^$jxC#j#SOXJԒ++w~n: W)l(JNk"'NfOd2"o9s]t777GiNv\ɥlx>ngޡ?QS"r숞u(D+p[@ S_jVqݬRy?rU)=I$H<)+" 5 "'E}Gx'ӝ=}:|]]'_fm>)B<.hMĒ;c:OZ#Q!R;U}U^ճmV azw?W}8_P $P8}x:&+;9{ T&L&o>lVG~36?fߏ=E;NJݸn/v/9l_9/6=^wwb9\?EqFmmy]LۛD>1?c; t䉯=O'566vy.7\P-f>AuZyt)}b<SLxvcwLx9,Ԏ"5o}n=bq{D{6.0N^lq"$|]PPrj̀auqĀ h_\s| , A(czfl 8 ?$Nd LU):_,'ӓ+k:hXomc:hw81ϴw|a/z v^Kg~>6 !ư&2G,:* ODd2Mt-%p.&l@e|.?;{8 4JĝHK dEDP0e=AVx8e_B]ȷ~1'bc2' ft%3y 2*Oyx_p64gJndD.,I% t"hE[GU' a(%Eфrd7V@G{o-yox_|5 ߊ|\A3j& a) eʂnIΐfyMo.)vx_l4LjıP2FYnek(xugf.&2Úb(8zEA.:C@>+ 3c s?n= X(lw=| :R[PP5ט"w;9 $9ĠdXVriQj ZқQ$f.ǨEV#=Fh9yE8LN2V+C Xj 2~wj-71 P UK80z+H u(tAm. 9eW|1GaK!w Bb? + k:h?kpvz{Ucdᛗ.Mki `_ H|^ŦοGݹLɋlo&u ZﯾlJw;]XcWv(Y]I= w a>6$K ;t=%5CY$5Ǜp6\W[^j_H%/ڤ:f*(j\,. p#C`Jy uԱe}!W96KnI!⒊)$e{)SR#kהAoa.ysxO)|b_tM|JkOp)4*._ vc^os;`y4FZ|ݥ8p4D V|@1MT3$ac9 XHÉ-a)xR$:eXE& &G $ Xrh-%-f+;WwǻneAc6-n}=Oh;c%!xu860C +.=qs|?t;c5ƭڢGjGzg8W3}r~Uni e#Ye5rxKk5&ućю+ 3htY?F&͕̄|q=2y3au2aB:a.y,  ®# 1]TB:Sr:JR x!`pi쯀\~,HcGc¨s%L`: "8f3X`%TvՕUr<.܋P9IMJ^rb0R]G(.́_Qi{)5|[2Z0\F7%5Htmӌϲ&HsoG')8u;^92v~2%>2񞟌z o_OF#2Er3\{J KuԱe}!ùE 6"A;eab8AN4j(j3*Jg6i oڂNX+5LqU< F*UQ>0J^j:E}s4w' BX+x09`QGcѶ{BUPU{,dh퐇*!CT]5-N)YI{Q",3Ib &OԄ+;zѿvo ޱ@£(E<-s'Ia/lԨfJ)(=<bKUqi$rJ8镻b^^ 222x*rdaL8λ*b!`BcC;6{mH/\[#Y"TfhJEǶ\$ h(`g`t(BEa[;F "0(_.ޞ*Yڢ_\2E$$5e6nkIH/C#JÁD  .Pd^cRYJTx/4%nu,/9Qqctq>.Wa U]1tX .jJ6tՖ X"WO92-sĀ'>go.-Wh xR^ikkQ0tf#Fzx_p{,L>tPx8ILȀorF\2s3 Q Pg8iDE5cQPb{,ŀ]OW ֗y*J|&0"@7QDą i jE-}Uj[B;3݆b|g7rTAEWAt̎ N(EZ/+:/ߤt |MsPg5B( :_t}Ϸ?s}q'SpIxRT=H7*%ToWbOPΥd" L0H(-Q&AL:=.(E}!>+9wލ>+T+W0:L^Qꖧ߿ut\3@ݢ͕UOJCÇ 2@c6*EpH/i˛II# AЮ0my":E:/f@7OoOF9qѪj 3wjDƻPblzh}]X&c2&Wtq]K#9r+0~|0v ؓaL*u$_oR#KR% $/wU!GI6R(^h$rSO9%O CmE&(Uhɐ_j7"*!fy'Yz\zԺ=YGV(yǟ7C~bB~Tŀ(yOMG>G )m4yRJcT0/Ka|+(Ar˜;IŻ]jz:Q@_jʫˡS.T@;!u`.\` }@p' 4Wn6z?)Vٍ]l{Oݳ+FxHuW7 -nŀ#{̌J&B l Mn6 RLu8 ?V 4} 0^bPqzӋڈowc{,SJ?⻄{GW r.R9uGz@??e ^ؾNy SͿ*M;xTMT@aGPgi3`i;.v" W\Mӳ!O*γ8eUpVŎ\%AOE+D'xZ޲6;=#cQ⩴.7Y.*IscU@5EA #"2I֬&N&x٣|EŠȌr"?srP_zh֚5! l[" 'W;eH;xЬqf !SO=Aig!ی:'ːyf][s|KhY"_뮂s uPA`h#ulk4l*0rs/z6&M9HO0GL *BT(I_L rKV3kƄ\^;"W\ԙrv.tY]"H/Es@`Dq2ȦV F$Ilat=P ;pKzfwϝ'uHp,]ok@b8#QfkPpl{IrSӫmlo,_r%uX?4[vsIsNX/Yͩ!+V/ìo׬!c7Cl,gy\t::xTX*AsF/>+/ʱ)A/Gr [ {Ľy2oG<^L/N'_]LQuiN\^Ǿ F D~%alqI_btؓμ8`vD~^>1&sk݊94X*N{6h5 zpFx;}eC!{{ 2DLO;/ $q)]a gCbj60jUD#ꑵut𨸃skV}تU~ƏԪ L7H]*,Ļ/ iQqVpsX'6ݡ\ .&?q94;s ܖsB! o&-[qa\Gѷ%=҃Q'<*yUAD{=A5Հc7'Bb8fыIǵD!FNk|]vQqH+Z)˻gM0eE<Fd'σPtNioTTABVg-dLOBL\z5EJE@! qHcsƄ!ǨM*_0wDj~ j4LtCFXGE NzpsM.ers|Eߗ_ֿO_3a|#r)6t2drٓlXdɭ4<45\A8\ԏ.\X| kTF t9/d衰$)e"F%[#(}Dco(gTa7(mrV7A})?iL'}ZV?lkjOu7]ti^>ZW!w|@et:o=fϽ0n!gx_0hZ_>s u%5Ejj<_Y7 Kdӂ +`7VZttXGN6q:uF'SR ̌>\>ic6H9<~\ryrgh/' 02??YsrcH;r`pE*k8u*h^IYΦI:'!\% g9j}x:ߧ+ΊbI iK q=2Jpr+Pq Hej1ģ ^װ|B<HEp6-[ZX| 9>9(}´Yr {/>Da$r>wt!_ X%xfs!&2z|cXkw]@B \^`<ɍvg?{Lwc*1Φ%:4 5N7Ʉ:[}`0IWJ&LDu6cc`‘#QpnIJJA\[y:Sѝ gcHZg{Fa1x4N9㿃]u6m=5cquVZM8,{2B6k?!I h#?~tgok{v,ocw;(RnoW>4~l;:wNU!^ .o1Z53cOe6!wi6>^*QH}uj>{w#m韫3qVgN5sv곙<ҩN_NS*阷)*Po5 0zoXb2="L(hǛ$5A +%a5)ApjJ;ߴwj"Vn{eyqf}[wCީ_j|`J8-$P@@U;$Y)$$1za<eӫcY<,vjho\_ewĽDVa ֛Y}0q SQ7HIqW~O*y }gcD{ ݫwx Poax Zr)8kY5޹ ~B fѨil 8g3X{ wx/!PnLI[)-h95RQD8 TpBZ|VG6̈IKÄd넙0& Gq <W>>`TxکSG˿wtZ 3@ckz +"0R!liY`;ݒ>[ByζP' UVS.IWpܩ^ '+[MVqw)CԦ,Q:&/YBY1nR~n!u"4G&icCbi,"6ʆx6r .XoXnsR9o rnkD zT>zs\c3ZY>9=8hB,aao`?'[]A DܥLjDcJV< i=J`Bcs(=/mCm񈷍R Y‹;m !(Tv~nR$YJOAA0L=v20ٹ)\nwv(Cf8R ;1=Aր!^]ܥ^ܰ}ĝ0+8C+c:b`lRT;|?~34MƮS]Hw5N q.!$"A7eʁ <8ېVfTmyuC+WYf^SDuݨr^{PsXr4$'8$E6(9<ۡSxЩ ĝΓ B8:Pj[j[ ]:;H[gkK˂] KɄ2Hig*0LnNL/ܧ+h~z<61l~Ps֊ e$4" ꡚB\_N5hMA5.֐!0c+4>s8a rIA6sNE#\8dŃ!0-(X:XT0=PIb*"Waviamr~XH1Ƭb6A󉝆Y6#";P|Y`>`M.(J~+LbVE@IzӨTo:rfc Op+|d%62V W`T!w_f2&04nS]5 6;Wql6wUd/?odEDBy:\? LQ -9\2ۧDKklJ\O.ʱbb7oq0b)=4Vϸ&UjS&e!*T)dF"vdzJwT:1P1zBQr<_K=,eQ$ 5~zGB>b;?Y; 9a{`g4esyOox-QGvfc5mX1M  !wx::d0HT҈sX)ze+@Hnft2v4ٽrNl=z 1:8ec4b| *2GN=)JL 05ԅ&^DeOeJgS+QmއNUMUSqs3N\A;وDLj,ˆ#t@٦#}qVnR;^w%:ȓ !aQ+vkuCԓnh7W ga˼Ђ_}~WL*jIKrL K:HS8 cR~vm;xIoĜT Д"ߤR¿w$7o7LrH?Apnjtbz4#ft֯iWW\$[F/V#uf|w e~\$n1[4urZwd4W>&uʵc91SgYjLN R+'*SL9ӎC`OlreiZc!zp #IS.˧c/H0ħ,6-yz+٠.s)@Yi7¸P.;./c{̘hG;{u(󴲍'լ=G{Cύ 1[S8 y&^x~9Z49wN,0`%0 +Sg _ Z'4fp4TC%)l0ʲ"=6v'wI=ItQlNTL=ݚ0B~ 1>*޵(\W;X|&Gŧ=l>fF5o1oNEJo[(wvy|9A%q!9~M?@?~q~K)]LPƕYp)S7ґ;ɽ_Ƙuմh-|/E!zmV=j|OϑHJ<\|pçp{ A"蛠q q3ZVhM}9Rk.AXWya9Q2D%#37aCfV&#LhiNr3BKge)-QM՚0<9b XSf96wg)kxR[̮[ !#}%\[Y0i}*`S)! ) r5Kj}U湗b0}l~,l,=Du=uSxg&2Er鮳us642̕ i|Rj>﨣#" 1ơ<4b&*g +n~hG 1D IP:QRZex0#hZL*{|&]G̍DFbȂ-#I4R4,/`I UϨ|ib₾~,&us6%˿`G=u1km;o6"SP _,ޞasWޟg7 _N\G=_]}KQiE0ݤ\Z_[ܯ:q~}~7ٴzwShV/"ie^2ʻꡟPeNwy7>㯶ˣG̼3Ŋ(4Wy-B"ubgӁ۬Y WS}sۘ_6]ԪsU"X xY1'w % WL(i/,*4wBXhptYR܁i&1Zҭާe滿Y}d2o7?\4i:r2԰nڏU5(wW>Yv[;}R͊Qr䡰:q-'3$Anu<4}'_Sf{xz!ik7%.0>u |_r~ CG#`i}BfZ eu _F\>o\HK=w!"z &g'+ 騣 1ʑQk]d3y#8QIHV THc[@Jcۊݠ2;Ubj}֓{]S/_;C$h#W!fB< j fBs-9/3q@(^nO>♓~KcK!v 9F= D4ㄔPp` ZS©s\B+nأbOAfYo`x" "ZMrR[ZC<@[z=2n71;Ezi̜J(8=r:%aB2[ %2*y= K ^2c]`wPuS;9A<"Rs0 &2"Ԥ@geJ)[f: ;KTV>`6km%ON gQyKеj[̓Ϗ ,cGNl1NkUci]mdZ70[.HDODKJtP#T Kh'YIx`g5|w]Y־E׾3:/-bI UoF+2כ6G#w~>u10S'tIdiA 'Д*YeܛC?P/`y0bӱ1iDB2ԗ#Zyl!QxQC7ͦ1^8")lQd6=|n=bO|\sjI UE_vF<&O ;VqW])9۽Ĭ)ܛ*JQG+fc PvWۓCϝLhw,Fd.;u"# 'jok>ަź1cPXW~eL!1 Ujnk⬿Ml8zA.FF'vӝ5u:J㢃ÜAi1aH!&V;ǶlݸdQdy5Y">F>1"nhk~U 1DU̫%9- T}eI6[suhG$ Jcc J.V€ȔiKK6 /Na&ilACaw tGrQ~[qO(`::0,L1212e5T|Y ꌲ–JbZ\;yC+)/]c¡˘tLCuD#`Y츂m.7Z" VhħK<@G&r:IQoB$ћb?rsggO^L\^Eol^xW /@v"Uo w0uta6F@Ox4S#>t8*E t)M5T| *t_\4v |rɯ/޿z.)]C {1ȟ5 τ2%%l8è1tBa-~0M F*[`?riˁ ~ t{LNZ4?nkeRhDŽݯ-Fp@1/iPϪ/0~ᙈ(PMzs)h0(e*")*vspww+:z`0H6XRNBy0ua6F2Zt98$Ol(vif/ۄx&H4m #ujmt n@B1*J0HU0#fKM.:T]~a/ec-b&Sfy,*k4:cRLSڟ`j֢W726d` j^8lhKZPNu -3>Ʋxc3hB(u. U*S)/r-ވe'{}/,5)dxJ6#KDaT\7`g]Gpc-{b>3*<%YK.cL;P,ڮˆQ|߰^*"V* lƫ̥LXh)e4)E/qK/4ⵀbzf/;}ƀGvYJRE=G(E٤Uc4:H9'~wqJYIXI)KGBD;2xRa{`ކIzu_Yˇj=r:EA }NzނOμz`E0W 0ްZ]vo<7Hw&SnK;jJᄎqۓg <: ˊW4;*?,QQ,3 2? ^fUKΕi{(*vA>rXX0^tNlBBm;` L^/nBlyԡJSVYtU+X|IT`A[nl}hm, h𹷆5Gr91b?!cyA~2MNGfPfA5nܦi[kfE" #188HZ7N !9ə)?W¶De֨qls%BϱWWq؞3$\HSaP[T83m O e'ñi#%$&rv&i=d@L(gӡ[VCum}[PKߛC+C#@ tW3cr!J4U0VER=?5"|(^\uLY),KHmu5,:8wF 1ĕzx$г68kޙu|rxF0e0x|3]ٍkpf @f9TaF^KP(رѼc43Gjp96XP@o @T<0CH Nd=sI='(ۦGbpb# 1;hĬQPZU4G"+JKT4; ӂ4Cl6=<#h6.:(M6(pR LiR |)]F ĢQY{LIvHfgTM>'MH l4ҒZf<~"/eICxÆtJQTYn-fZ`i%Q;*E!!>'ȯ#18yrTrDՑ}iOa6vv8t_BP՚f'ŀ0KGbpr~d#:@V^ŕFdZg qi1p0rMخy$OIK3 ٽdY YS&hR䊢tL}5Du_X?Y3ii$o {K̡F qٚqtבixPA=FT ()=_L,x11tntS"0RMoz p>%H 8kו2r#u`F.]‰a<6\Ře z߰3[{x$;,f1'Ibx"t? *JPr-?"(E61 (#18DF2|ڣY=`ZGPaѫh{Yqb|yA}Fw Rwe3 #)Gbp$;抟q#B#7=6i7EIv:q$ Г, ٸYjwӀt^0 D=C#18&mr2'&{nN w_(hQ%6~j0Qjʒ 5HL><&(C41FAmǑ,oC #18<Axne!rкӭcj} -f[bȄ+)5V3-̆3$ךȲ]߅(K,8J~5p9 6ԍ1J~x#.-} _deYY."Ԧ GABIneN ,PGQSL+. lQO HAqDqR l A TxےY"df7كc*bnxIE nr%H%p+|=Y4oK",tȾ8L?1.,B2}FpCҁQd k*T~Rj(qOK` lȹ/{a`Bw>߷]{]L.1jJuaִ;u2<& ĨD(9<ųMf- *G<1cܦAܦ"Ձ H BcrLf *9Fř/r4_B7P!_|cT H NoUhO34 ~#>T A6" M>i{}x%+Bar iDH@pف}$I?r'aA*K,VW$"2;OE]H^OG=FY<,nhA$Set^?##XHCGA ͣx$ͣE7+*CB(,BvGyc|n@ڠqy=/zV͗Q oE)ғX/L`i/vˆ7\Xϫ٠q?^ 4e?VcA]Kc.3ؐL)*xe9Lp=fԚ2'NSKyf~~s;/+7o_75ȡi),B [75llڏpk'෇v2@<>0P̫@w^­,y?Y[`!M^ԣ뫃D6`yگFBy)M__c"2W@qV{BIS.&ۗ;+} u~۟`);~z,^L^wLV/S{ZN,cVZre?ZBxa N{7p5qp3V`ɊM\yXv"=?L y2]#'TAg>+#ௐ#"0 `xA\?.ݘH@~+ r.B[(7_ yzr0s\ik2X!P 1wqJYIXI)KGBDpn9x"{mm[P_U5`hvw/Ïk"W+WoXZ7fMI .טE2Սԣ IDb{URwW8Rt٣:%zx^%p^ek_%zxatؗ|O7R/7a.$EKоruq|hTKmuG$vGQ#[V"aKk_mLp;! 1n1gaw=nH\d!ldn2A-ZIv[W-]uu--v{dTX$( 'A qSV';_.p$KM Z{#@7hd"2q[.8ߋl YTGl3oGڿ1nuNI57:x}DEAEn\:SD!~{UpV )_G=j\_Dߴdب?}/ yK#!^.^TmeC؟3eEo uG\9zTt5>^H> Z5!s&bbIbss ygLXNwq1Ș](&B́\a<5#JÈLMgv#ρ&N$fvhj[﬙zȓ 퉵^kToc̤af?FLdWkKsr]A ?;,n:5UG·vhLtوr2f\rbnlsVϠN~o5{ٿ7qaw슍QH3)`fE6n0Xa_"n~yW,^?YWo c)╏kr`}{c@z =w_8[r}b7/Q(?$^Ls~I 8Y; Gܪ7W|{{ѹSyn}<2L__N]-ynw0U01'yr:^Gzib7R.3#-Bas߶i6 ٥8R@x=!&q>{vf/۾}@5l{__8tTG7^T4?ŕj/>;e2Uf/IlG#}vH\Pk]ݺ*Ň<ڀķUTG^fɨ]LQ.ë!Mb/m)5?nwzD}`yAGV1RnE_"/r"Z^%Y×.JicJM=Z1hn0'L)]r!ڷh#ň*N}&VW}.7y^nA,s]6)Mz$s$Kqg52r8jC|ҎYN#SxDj-o=F$ ui%#CLTLyi۸_4*0v7C+{8>^yI_ W .*3^=#W 4+(Kl{ڸՃC9+:&~{]ߙٻy-kª0Km'={uS SYOm%v=~gԳQuy=f=yț\4wwd![FG*I0Sxө?""~naʹ*~vQ̆ ֨?ҰEf}fy𕶀hꯋ^˸h2m܄.!mmL| w@AYx'tdzpPk5L9 83 ǯn pQ߰f[MATpR}o9{]p:R=ũIvaћ*¢0nYkU_aC1uhGR^HZ9%)OAj(QyH,ۘK1<ナtlŻa (ٰd%S5yR'r,V1KtOͣ)+Vtq^ӽZ_X:n@.Fz'k!Edž]V{Dj^gRWkncuw2ݧ4%-x#-mkߥ{1sq{"kUG6[2} f\=Ejށ.5x5<1qYc ǴQ>T<-zW>"R""bPrϮ)1gov}^r(iu9wra5yK>6'E^G#x$۟o{aۂ<1R^ !.'BD"` ryJ#҃=bX\NoB=,@7޾|{Wj1;s'Ct/S.yEv0&="D}r Cՙgq3ɷ`.E$RtCLyXḆ:f>]BN<包\%X!jgՎp8\ayHHD"2 H^w/P+ykZyă:P#G B%QLL S _mUҘjWJI(M6g,F0W!&,0x%j6x"EL(j9CGy 5kHN)q┸x8xqGx|=.Z'Z+nùp@.zZ=uz~LЃ|!⻗ P5̵6JBZ+ e: GrBV Hkc a 5~Z5Q0q,DH0x"Wgr*MU}7Wo]>+ZqӿP>N&e|b,ޝmPDHCԘ16#.l[+];vZiHۃ!aMFC)R <ذBWЋha i[&e ی^Fˊy:-1xXmǷ18e"YtfBY)!x'2Y$^4~Q 23$Q7ja.S0 UX0;nI`iO'̷HN!ZD|BdI0KmК3LI\,9hbE' j=.Svw_di|Sp,5gf]Ѣ|){U@>^פˑ_VOB{I%a_27*Tfcg$rD3WbvEtX#9KIyn]H2OxlTEHψ9Yki Қ <8lphE ڻop]a(]5޷ke7ful#4;+u +Қs郓Z)J!mJy0$hvWp'm< J1˔a}s9[5V#ۃĬ%#0-j#NF"~`mTH:olƊ2' A*Bj.VR9CN3Gmu [("asyAo6z`#`Ё<6c<6c<6c<60KUc ylylylyl"D:`v2xsonm,e"Y2jD.^V~T̻K?mpYM+JA&c1\_Oi0TZ@-HϠpiK Ep3%5"f"w&ޑ\;\ vٰ&iZ5I= zUL˾_.2+2EY2’R[hr*~υKv($u1Hk.~9y+m>&HyZ8ZI$pNo_{=~'..$ q!,@ /6rC2XWz[o&+SY%].Z p\`kiA$ԟ;^`NNɕ ̲Eh= hv~;E&Mhic4m9dvb d{l+dբM&!nYJv4Kd!SkwvTVm|JCsĖ]RDQ+Y5l;vkUɫu8wO܉I\5X6KnR66/]1MRZF(#6.J p>BnoqZdžN|(YqE/-Ƈ!@ܞWh1TKLV .[׹M |Bxhȕ`lu˙ެK7d,t>/n4e."JEw ߱jVNjGJ5_F"uhɋX-XՒ9އG6vY:/)ę!T~-Ic՝82> yWheO)X{hU#Z:MnM>;61<run-=on?A.p!Sj!ĴXFޱ{i 2WI0ީw m|uX1:sO>br,si0!ƥEOsω[IOؑ?lF) Ŵ Z`ٗO~b0{?[_vm0-5؅/W(d_ƭ~Wϋ3]74X^Y};Xo24`('M[_0 2 .Z0>1 -}}Ud4dp8Nff57Mf TVfƫYx ʹS p鋚=,)1Y@GK7*{tP|.U "Fb"$L'OQL j ֯?~Jc@=|f-k0o•zw~7XG|B3J.pj%.ܗfb*ՠwW -eZ]$dr37`{u/+b8\+saqKύ2DK_1nH^<K& *';+6~} YR_[aNN?(.h ?͵8y\ RՍX+uOnVBjPEJ9 -}Cjm#7',[mv;}m6wxzENj$82QFɧtVIcrNumpMF$8/S[%:Y+]SIȰoz·Q5t#IѤ/M̂/ &VD~(Hvu¤߁OOF#ۤla8.stK:i)hO+\.onr (Dr]ont4I5w>+Sz}^Vs:WqBgd 9VEf-QŃ#M4Xj!RXjKmcm,Zވ7ZXkyc-o}PJRfpyX,cp,b6M]YؓNFȟtm8eeԗJ+* -3?I]u<3[eT[~SݠX<-ͦOY㌔}M޶<w͞L(ӯn$֖m]'cf4STnɭ0vijl0zC‡>aE OC:S'=G,J[DճP"{K||W˱oLWCt_qsG}VеgkPxyT;v`Z?SQ=ڻGj<)5O kמw1*{S{HY:z~gkNO>T:ݞVrž@[6Eg[Nz?|i݄ƒgkøy uT^e2罛\+y`7LjjƬS뵧샄OV]zWzbe{ K@߇},MfT 5ՆOy6=nI!ߏw-/%Ţ3:kY_Hi7e%Qn$enֻNͼoNtr9@ϟYqu)qgc/'.VV7{̤{C&/%yr+c!sOU;)Y|߮ F[>]wqV0{Ƹm.ZUU7¾gKz*6Uf4A j˦!r%]ʐ/GW \rx?(gCskWmo$=Y:Iqs^GQq:1<4RhVnj:`0Qp1RIjs}=RKS#xm0pk&[J~ 8Cu[r13<Іd #ʃB*<АkH#T*Mxנg):y]luF4-+\v~cﴫ 1)bvHE`]$w)Š#jt) lۉdo0s6p779)#+si[bos XQI@WzE,>[L҃c֎JB*단d{VuFZ:|>ksP+@:X;NPfa5xA)-tM=lwMۺxvuv6Ӻ sO:e:R4=X*,>>=G٢xPn<:h@7x{x|p7 ZY ~:i[ıo5VHojUg\)HYM (%$A䁵{k% ~8}?Y+ sE]I8U; kID@2 YW/^: 4a֝2M5cf;x@HtD2o.tƂ0leoH#(|F,g|F,g|F$72rOX^ܬ }`Uc9UHV !pAЅVO DUmdM|)c Yҍ{.'y! `R՞GZq(`#_{_KׅGwkԃ^\_x{M~v.q,M2v􂂛On®zT xCA b,4ղEیݼRAT g05VQΘbKa--.ȴ @Fqc`79*,c'u4>]`NXϻ#3xX-j5x-,|v. Lrd>\$8 (:EP! -4ށ1E! <*=5<@b~iw}qt7i,uJƽ)%X"5Wk5܇eiE&m$9^x:}.4|ln^`{Z{OsDub㲙pub^Ŕ 0UըV}s{+[]QStjN;aMܰS=<﹭צe#֠;6ɃSn~tA4㣟?v O/?>]TO_; txcsQ6Bᆭ/{x0?̦w޹ Ϸ.&5#nvoXsG[GI.M9U_u^OWX )9 %I}kZ71j dc4d Ioы'B~IEbOR!}<}W ר4rzM {ݪEsHRԟfs7_=1=!$@A=%i*@*NY< ڡA>Wk^Q#`⛼?Jӵ۳s>TX/m:^RHQ[LFehv *g5t:h8ơn}M}񂰠grANl]Pe2vDZV(daᐄ;vA16ڛ@MM"%"U'ۙʵqo]@E6O@SsPNJ=L>% X$󗶠뇊[K)Co B3^C@o$Ւ,db0$IL`tWlpMlp%qҤ.E3j(-f,gt7F; ~Q:t lo ǩ:OsOʬZs"{)/?X 6t3F "T+g!uF ttzum[Axj}h:%KIj<~emqauB+3DCtv_v- ى$9i9)*=v\* |ԭ0}t$CN!$iqy-D*O?!_za 64\jg*Nt㗒?V3ƭҊ%/oْ'Qaz$Ż] I欋ch ϙ[@oA5:?q[ 6KID) },Z&m繭#spjlLPI'dLȘŝ˧vS4S{Qp|V(QFa~/9{kCIJ] 7/6PNؕ+r&8{{ ]}m#f s&`9$:N$C'PFQWk5ѕ4 s֨1/U D=Qq@p9<[Udf\7@fW=LFf>\^3|(J5{peJd~O? q9o?RU~2 I q TH5] A+gܤ̺YҊѫxV<s~y?%(8ju^&~wv# np0_˸矋7Bc5/d~ ׼B~7?sy2|mzhV',~?W3 ?O';rwGXjO]ܵ/ `@ ~|Q_.}Su0u85+Ex QvV D[!t*va-bkYI5!^b2?y5ՍH8y&=ZV2R]sGAm|[?A\\; V ^v0㩉 >=ɠ ESI6&:{7: yo2khɇӤIݘf_Fy%&/nIIy^~ 7OhSs0+ W5;]7K~^G)coTtq<^_' 1e2^ߚǟGsW?_K݂Qf2ualO v70ݙ |rq:0Xn swl0j.w17BYfݲ|;< uR yTMhn6_ȏH>S AQxAx|!_\9x 7M' 6~]ZQ { <-]:Q'Ok*_A):Vrt,GtݕOu,F&rr鍳]ENPjerj;rl*Z9aZ &x,4s ),,,ܬ0h@7x{x|YC(ވ|#q]jNnnw MzPi \~8\\;c#P EKHk%{ Jb%o5b'5:{^P+I^4 ^A&sj5|>h2*L'K$Ӑ~8)߬S@,XVs邏^of_}̹2(P_ c>#\ilv&yS9י1^k /rFAR#ŝea! + ڭ,]-k5nUp6*qu@X`Ob4("FE7hF,GrE GD:"7E:`cA@7$a2c Ipe v<u00, S LlW d2vC2&qTG 3(Hj#cA0&p8$iC=A,iYҲ^KXݰ6N7D< .i]7c /$[lɋdKRȥ|gxHmS~ Ȯ?K^(t<'Sq,BgZ:6:9Yh1 QQۿ9$_.!e`az>evLm7|perפ6BYekΝ@ךw $ET7K>a f1j_?0BmL7]2lZnlvJu2h m'5Ę3ؤX}90Pɼӭ0q]x'Z\[d-yQNx; t6\NX˖n sUo6,W3/90 j}EUo%oU2\ jo;%i3/Fv!:1G`>{2P/cD&K]T؉z-I *NR^;0ٻ8n,WشMˇc3&|BlВx=nud$K~A`IU]qyIVvɖ"e5nƘ/ЗY-b  >Uղޛw$e9RNNʆ(ӡorȫ_&`_Ϛ\KސV/?;u=㹏,{;'/ YRoZ5ZY)P{!*g/1j%pOݨE}'O7 PW5HL2$%T8 c#GĨZ,QŹˢ!._(wpϭxr8ܞpػAٛ Ӟ󐻷oNW!/ TjѷWWVʑ02Yܷ b'«;y3{p\| y.Yhd2yϙ8OyDvG?:ۻeC5  077.Z(TbRLy_bQ(sr=ؙğX;@++lJё:^E o>[& O6Kx&GtiQPvβ~z2Wte̙g˓SP\O%iL5s&U ϳ)IPlb>Aҩ)o_?TƠvƨVu IԍR}"He3jώ^w E 84m7(ufqv\L!xw ~GZ1MC,b9G`$w P͆#҇OFl9Og=ޘϋ"엽e/~ً_a /ȥ}+iֱҹiqb:[6ZLL1Fvޚ}8}7a X TQfog;D}jk˯X=',u] Jrj_0_ξI CZ= }PP?ӷ̖գ?#sm4ju EmSĺ5/{{ٰs<>.S]Vxis<;CyZjqvgDLgb_͋`+]c! ?;V}:8\ |Åa@ џ+0pqϮ8p/+%omx`km!piaexEG]ܝ@!^0g+>rG {g̬m7ydk݈ QA#Aug wa}d/ݟpmk;!/kctxcK35g[ijc tg?hwЭwdz;dR(I"0]:(.[ )dv csTA4kDUj*[6Xs%- ߫o0@xCx nH4w~(zFe[Wgw:`S"XGxkˮh& ;1ajEM-8eFZ=c=bLC:]ڵ;}+z~kB{kr/2|q{?tdZ,Mq93NM zLm9)GG9K>T胞2tk-R..J-Rr6R)SPCť^ڏVY lNjO:*#58r Fj1PBN ͕F:8 Ôu~OK;ܷR+`pi+"&7P<MTzѡhxxQQ#_K0Rli~<>1t j[OZWˠRevƤ.j|\qQ}j֘݀WW5m^G|nV|3{{+G'‹y1@;W7 ~#W4lv6"o}I{g@ԓWhT}7:KU;E߽iŚEx3 }odyW_{1(|i&@q)hnvZ3/zz<5(5HGNjY_B7+=2⇟ j^.h6|b|̷t$ۄs]vcdeRѰo>G"#^KXb)%շ$f}6H[ "{,t ،$PʦH.gU}h9Ũ+k{pEk36Keɺ84d6mW5P&RCW!(x]|zrEGëbiͻCsEd7,!-ZW#T{7qK=;ysd坏Y,lC V%Cb}J2X w"UY;USckJ̱6'TE*Qjj"ȵS4Jgdn:|_~x8ŎI@-Ÿ S lʈjH)B"PU%BB}"jK566P*<*,I[M^CUq QM٪\ -6r%d1"frYX ]I5n) ;*X{R]J}f= TFE|pe@@czarJa%D[1@Q lxUZS|kålg#/"7x Qj<تdKW%^dD6|) 61ad4t2 ͢c JʞT}``Mʣ8$im"RR^6/@&rrq!{X GW ZEBpl<&,A1x7 zB܂^rB*R5#b*J C26#mM}h`ǓÔNjETC $<`. kNC|+!PO),H,`@ђ&Bx6nYQY'ys3 5%AG$,I j(*?%svdr=I0΄="XVUr ֐(4m_d gT3R@Hqps$eQaj` ',+`HϞKCR`@H!ؕMH)?VvOJ2"z_\l4 FT} J {T~9qEH :BDBRX,KƆ^}vM0:o3X+꼵 Fǁ xN-f}rn-F v }Ĭq@r"`{mPBL &X(y_7S n Ȗ ;5~5MQ ·k(/]h@Ӧ(Q)prt9|$\:쵠p1[$;ƅ^dΓ `2A-)J# I Vdx#p0¸GƁ>SA^|D$H.:/u$@x3 nCf6M‚չI,TG7gojDU6qf[&K ^V3HlDž56}mUGUEq>e[5}6Fx4a=RA"Q)@r!aZ& 75ڽAX1z,! 6#-h'Q yYz6^|J;hA&Ō=ZMVXyptAGda⪏t& 3%VYds ` ɃX0"b 8\-F ,0n**sUB@:C!R+Y<~ L}tRBcX# `f9v୛'s)ܴ*^ `nUNx&( 2 H%/V6{TR%4.C?wyO:oz0M{z8k;3BpB`&M E=*588B?P|wTuZ*sBk9GբFK72(#460caAMr9VSFAt0 i_%}5  )7 nVz K6w}ύF_ae+ _\$UeK?hFY>R1![ Y"ZUlv'}M*(cˮ( Xqu5 pZq*ZqY `-0{VfX ˛ay3,o͐'ݠ#RƐ'6pPɨ~_=Yp\kуWq/Ő=LFe4W#:>6Yuo1*G%yYB}s5ʒ籯 IE>E>ڤodnj{7K w-}6ʏl/- 9 vz8F_5??eÅE3ÇKϘ[C{Uz)PBaUz)Er.PyRY&Br5ZHo\}9Nr#W`j\j$W#HFr5\j$W#HFr5\j$W#HFr5\j$W#HFr5\j$W#HFr5/\OFfaߵHj0!Wp3^ :hJQ*\ՌSCOWv6*AZDY"ZZ!  nv,]yJB* "5!XPdWvu`]/zWuȤ@eDYVHH CS/<2d\wJ+^wWIk@w%b', a^!+y0¼BW a^!+y0¼BW a^!+y0¼BW a^!+y0¼BW a^!+y0¼BW a^!+y0zy:Bv=ySv5ye}y;pyB^M& L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&b=7>c&?|(ci[+ ~ۅgu}R9bWU%vEĥ)`R#q-D7Rd>|w?bCT [f yx獳ңV0LvB! z6Lä$5m߷6Ds;_9ۻ6d,1K QbX.1psɁ9a9ouQm|็B RK4Usb9ˉJ)#=z[<q.!:ʫ=͆wO̽R8]:p=yT\Ki zƊg< *uO 6) lޗC}/B;Na퉨&_ -7 >kqی}M~ګ,^ Ʌ* erM)y6NF,*JQyQϨyʑz]MZ uz0[Qlߜɿ08#%L~g2(´D[Fhr ^1^!zRgY4'p5pcS~ʁzܳ4hDGƹSI8K ANߊPa0*~ _G u۔gսŨeUws9*j,W I;P&shl>$w$W;ܻy.ٵ|(?--PpSж`*f;(@*|G5T6¢`|i2Ƿ'3dMo~oa+#; =?J⑛-60'_anit)(}A71!2{m]׋kCB,x-vW Nx<6ayVAn{ VЮ=7΄4b@D&3f$lت2ؚ?9|D*_p *y5!.G̨ŋ Mmk}(G>&TF|MD/}c#PJ<2g6CIMEs[ ?ncrr@zrebwkuwo;YKG RH2fAvNjm5^MƁ4pBr?yZ5]ȾyԻO?f2*3Z= I@(f.0"8dR5e017x 6ly1]L(|8.Z;-xQXԬ8% a@M\-~T- 8O3;jtuRqTGa%oû=2CC^x?9'o ݊etŽCIOOGͻے!Ҋb +b0͊[6"[wbF7OLRД7aptJ48:է18)A*c+[K_ {{#RIKsۧamZ-6"'Ovku0} >K= G*M-TS:X şRo-NJ :T0WiW/l} T36#xEraI9BiO*P*@9*1Эb(ݴ p|bIK Sܒjy 1Xor y^6~!O/fBTGul Z=UFk,ia;fetxVD|9Ѭ=z2x(zwӁE74 QYX]{-v2ŞJgfn>uO~݀)P'м0s(q2D,1"զ`aTˇLl8}EVHj `LS2@biւ Y]X {vĂ::tRCU,Be)`ET$"ҌY&2gpK:qhvJMj ;R XP*Ìξ-̕cD9i;4+g{\os7]tQ6ix!5L%cV{c1j15,qmǃy`ŋegX2qCD_f6uedXo( EwF߀^Lh'E]7}XC|Ѥ ɱ# {.Po5v8[7 }6u6yz]xw> FPy+-D?%ɼnAaꞔ2pC  ,0tu{ w L黍}&%`ܧbw 2(RR&grطjfۋ/W=\cl:Ʒù۱'oΝN]ޙ3,Q2&OE.׹)lHmE,cL:sPA nc\c + G.;v8^'h茞հ蔡Sv,y޶OvȌ:d,yk-D#2Ly ;+NEMF":2̔tr]s X`^l6Õkhr9],yܛAu )\"/B&Yfƅz=2JZ=XOr`F9Yo) L}/0=I{QMmkO;^wv xɓ.4>>*Ψ,QDǑYHxRB{E̺*_q)It2$`C2\謏@FQȕrlSudO!1"*%K$R^>2&drR$e.R?rUf9a cR(s, 4p-xk& d阸WfՆ\QH@1mȚ%\x5՝E7zg9Wce7%nc#sG"s\N{5wI۴l iʂ|s&l-av'‚R*<9$ 鳧绡[ux =nލo˳wgmy:;Tt5x( ?^=ݦ+ ,=OWWI9l߫O2 Wĩ%tIc:[KuZkJ:I;NkCSmӀaT |UG:-IAJS8˒6BL(93 YD2%Fg @\[6Mַf9OΟ қO[ma;ԓڪ8uQ< p^#j7"|$` 6'JTYH9wmQQyasL1Zs!a֞G-eXAgiQoV !]֥Htld yf1!s5p4WVf*jh/ĹZkcRy|/-!_lGGZ IkgkK4S*!LVdpY1 ^2H ҉ArQs΃JtR(pr>{.yp3ffYUxSr))fS< 1)`ȑ9V8+B*jٲ L7_B d}o~[oQT}s7YˇY: 745+vښwpfvíVx[??7[! QʹôT5n~$nnvo>K/գ_,A$扡HN/#Qdel̽e7 2.J SP@s[)gԍF U.~Gfz:qp!N=3 |cş<\2wO@sI*g(i^ $3X oᝣ qXΟfJz:v/9\ 6g.إWm77J0*5?Be ϥOm'D+Qjh~>`M j~'؍O# t"AQOЌ^MȲZ]D\UI!joѮ-jՅvT8fWK#U_~dEKKN-Bq&[<(h)BxRu1r ec]VVBz٧W|VǨam{]+k—`tfeƶ!#!-+؞\Yy7ζ+PM' V&GFLȖԒ6BaxX I7)5E pYWg/D1=/̧wY9ZYNoUMR;@NVL/W9 ޲ޗ3oCE%,ɧC| R.@: BdaRAȨbN(%4PcJ=GvA-VyP(J?%+x`!G\HkMMp6hU2/3^fs4QUZ**s{Zʵ*cb$PY꘭ѥhr:I1%c\F 5⥘&][JR 䯢zbqg'y'vq@OS#2 @vS=AW5QgRLj@$aY*7R4 4 SL4( m- !p>fi@s5$M&cM.HƬa* PlV"H>` ɄpN3'A ٖm$I>%%lBQ6YlYupTO~qN4;zvkFĂ^eI BG%22$yH-.F,+Sh0 (€ nnBȻ@G߾O짃Kʨ}Fx5Dr5. }#/H&xQQ)& 3,1 *3nw3"1 8u0/ί1}Z\f5BEUeiO 8=ryeMnܜ@& ߹/[1pE燛r.E͵GTJ a^5<\NwW$ f n#WG&&?eԳ^5N1&Z6\>~m Wdz~ϷAeRK2Z}}(j}>` KV/EMo](^(1;p_~9^iVVE꠭ٵ.WNwcĪ)nq}< 3Ru* ߶_s&_t;| 8<Dn^4[2IcNj("Z56]DlR':'c*Bf/4v#/RF/cĹq2R8)Y)'nMoX*&o-씷BYCpb)d r4dϵi߲͸3d-1f8ܞᨉ2'V*iJ)hp=nE.3Rif198 J\[e?Il]nh;\r2 Hm Jy2Р"=[$ts7_Og082yR 0iO3\0jI|)[x7\{fl_S K-cgTI#ȅb;#QAR<'QA#ϑR`3ɪ<2:#>PhGpI})Uk;&,{˻׻cEF񺀠=L4:{P̬R4q'eg<)>7 Wsqړw(%y'Sc{wuzC(fkN&L3P8H|tŞNHw~RP*X%B I w>jcd- n#}5=p`}ϟb.ې5%ح\x>SBYڅ ZT 9$sŧϹs+yuFj9%:`T?Mf[O8\s\2 q2Qv|'AȜmdݯ^I\ro1>QHh(@&A Ng.ד' h$VK lrpq]8]7rMmfugk) 㴝ɻ:#C rœFh\_Rfܤgg6uB1 ~t5x(4ë ^WjiK2c e[Eu&gq8ByN'_x^3%qx 3o<|Ǜ|O8og;4MƻηQGXs롇'?Vm]s=ꟿAV;d}se|sF4}EI7hr%: $p]}clHM>gS dmFzT3 >WcTOMK`Zh*Ѭge6mEZBZ Iμ rYkQ) +ʪ#Ǡp'wl66>:n0`j!C٘ux}g@<:wGPn;4yA6&gۋzٮP?orxU$_}g~&[Z|}m&Иr'sC1VE!ER Hqe[Q#(H߂BS(z@)E/ݚ9E4̖g׳ɝnjbS0o'[?7?bjB ~{dݫE_YxU !9G|uH2 n*$X1H*TM?| ^Lyk+d#+;n_>_ hh49^uy0}1β^@S%8 A1TBlO(20L>&Ƴq߫VIXRYﴠ)G"ED[B IGem {2VeɘєjIRQ5D%U*!c!*iUҪSz-agYP,О{MP-xFC ;5P" ' U5 h,I3ףּX3Y?jً&OGhrsRff:_ٓCmz>m Z@%~۴rrU2Hf*ϺJh\}`Uxcx UxۆO^%N6LjzI- 8QYږqYAOm)I{YvJV /5[Gjm.\Oflg+bmf;$w:t ICp3欔Mڂ7 ׿fjhDψ_B'3\_"LQә\5GJ*jZ !Ͻ_T]!G]er)2u*S+J AH]e鍺Br !}QWZz*Sj#+- SG F]erioUkWWʍ8sUWG]Pi>z"M&?K˜Iˆ O軰 ؗؠ04.:Հ ,&2C9 |k\]^vaN ĦDR$KN[+u2bD%@RD(F'l5o-׼嚷\kr[y5o-׼嚷\k=K O[HٲBogj35 =LS<DoҔx,[NLPi6EI5h3EĕG"^"^*"^&0d|tA2J$Zx >Sb9ka-+' IMu"(: o %8j>y{^u׽ `k}|KyرݽoY(++"@jպ Q Už9{j-y1H:L F\D+x J`A"wOEb!k֔z.hCL$ygLD˗ CMG%@ *$Y(ך!|2 5 ;^ؿAqGD0hs/mW_8?kׯ'Zπ6rxj4օz0lG\k.}QX*tIVa6 OT!.{'| ?spG׫_'?2r~+XOy4g\j#FzQGԆ݂SMiT B)*D(aKCyIGh"uqfhDPƛ.o  S|O-$9I$pGH`,Iƈ $j"ӲlB8C`uݠq[ovy'f!h 2(ZV¸UR ^x0Q$0<$ om;Ovv2?283B[|[N&!o3q tgK50^MP‚7aZ,q5$o'=q`W>ĻLq'C Aznl $7 Mnt IQK9?LIExb5D..Y xy`^F#hIR^ٵ@ȶA1ƼqzBFK0=.9) 6:iN5(T1HL451[DA)PpkeaxY dt{(=^ v7Pu2_4'LP |=;pvœfԭ>}tmēww2I:ML]3:M?!Sx8ϣrG&}s[ \N#ٴz̏O9=a捑!׸o}p O8g-l'?2q[y'Q|;Ofv5׏zc?ѭ[(_ieE ϭ)%ȟM,A\r5,Alj%?JvM]zb ZE]{eJ<SZ"c0KnS6 =:n0`(&O-9 5OO0ag7J*_x &R{%$, wZP@M' = )D%L{/Xģ%$IȯXw$cFSBQ&IEyPR$eTG@q(\uѫUIvNI뵄A( &"߳4CY5Bo@&I&7rYJ.׭Y$N8ibEcI2yoZf;l,Nyl2 ~Ԍ&77Y!.ekoɯ=?v`D3[܆A`K>^Jǻ[0`)ٻ6$W?b>3m4v_X4$MR,%Iɣ4`YJ&#"3=cbDml2j$FY$:`鱶6aSGdm: gwsӷ.mJ_/\,jnN vm\sd%YVY<8 Wp#Lvtې2l~qJT<_6r#Y2sͲ.NX)C'UŴaK﫽JJw$Z6ne'-!2Ov/ڌGUʝ[;yY[.)m"nsftz?WkTQ gF;H`<mYq6<0I‡gmΠM؏yWtMڼxY=Y]n>xnu&#¦Ugsf)dDBJEVR:w(F. }im 3"$cؐPV6zݿ2ҫS%]p\]pˎY˴&c6DL,IE)1AbX'- ]-(]m Jo`09Kv#ʶ" -J]J ּ3)3v%?P8*5I`6萸KqU+*Sk+- *ސ2\ $7 ;Ͳv5.oɛ!_9lY [o4O` .NڴQ?i8O .õ|i|y\UZ*9}La\r=Gx0_.@ПZC&0_[kz;Y6Bu< h_)d;"ϩQ$` ??~7M@m>CLcͻO?{܂ -tNܝsL.9y}L%ULA1-VwH\I Ҫ;2+dWUV^R*ޠR%t ֺ3J1cgj9wqd7(4\+ XLnwUVsW@U\EqetH\t8Mejٟ]e*TlWoH\ "Ľ>07}\+):_}z[&S H a,r[]ڈ%S )\ۡh49n |WأԂḋ&usO ^E@_8۔19lg4X{,(!/I%J #F%*&uF*N8=gU9y}k]̅g"k"<R*4!Mr:*'&[^&Z$>HvErijPfh> {3jk>iK-l -K`6b9έVITH,E%$ J9Ix3JDʦ9䴳FFC4V[\$.iDDKTx,0UbGq+h4Kr޿,>-=Gky λ6ż*fΟ&ph.`Y6e:^4k-uLo POeDSF:jBȄgnB3!,LHl7 d|ۑ,J`:Y (ʰH"ZXKdI DREI@_t|`9DnL5iFPMh1r)sysyKY͡zr §ᅩ0a*g}Z1ic ԼYvܫGB\RʊyBj8PI.`t1{TK`Yp#+՘ [a~>w4Q0`Qj/.( J"T, WV=39s#Poc@ҔoߠJ+ eTa8ߊ[0IhZn4 ]@`3dRHzRQԻz7zw*c&*J UѠ&29KB[8Drt7& 뙱1$ˆɵO3 hs R%ft b>p V$+9q'j[ۦ(GN4k,߾ԋ#\l¶\bb2).ݵ-7"i x,Xse4E6IBy5zFoi R3c,\8O hɭ0yOKNJ.'V3!E᥊(Ek18_u@!ЄDX 5$`Y^#g譹l~ [9|mInͫLOx.t?7?٣ [5V h۔ {r~X[{?}hvol;q?ty@ߐQ7;bcU9,|`>1ҡ [ܪm~"ۜOiSH|9D9tMIǞF%9 IHM:Is0rSu+p,4hrZ$ ~FzX#N8lJ<Éad 3a(M 8j"ղe-Q<1C*&HȵݸBqڄQg4YN/}Y[ݧi A; +c}6ܝp^WO4|l3sJA>Q˃vccL) Ũ(Sf$(7{ئʼE56_E;mdPNJpZA+~zsɥdgyL ⃡*`%,F"5Cbu?qSδTb$]h:k h/i}uXX8DL!ODO,hn0K;oN*Y'|p~nC?3 dhطӶ|PDP( @3=Kst.Lˆ1P@ƆʼnMSkwSxUU/۰jZc2e0t,ѲH4H19p/EYr(Ɵ>}gw~ҦM@ąM~([ϧ(Ϻ|?Q7/_"B7HCwyNN` 4aLw|>g]v',٧gzGb);|xS1$Xb:f28L2csRE O7XtDSX"i  RO|FQ&jD:RX JE*w+n1(dǏ?5Ʉ f(IC:ST*H۔`[Vd0c|*q(!\`8*0숑E%*مBj#DOP 8l :%n9%"M JKb얌QIda,Yؔ(+ M+f7@yClIcSy|  Fl,/9k8PM3(fQFD%wEMP ,1 rDŽꔰ[:mU90Q KB*90 E"FL IXWcF$aAq2#QFI{+f>;\ *UN;$uSv:cIyETw0BLywRֻ_^ '(L A"0usG^Saj"0݆Zbw)J%' ece+g>™1+c1,ʡČULq0Ty^^ULګɪtQORas&pQ:q" i2.taY9I#*WtR&B6׊n`FG?qV},_q(ںptƓזיmԺ-|x6L^p ڄ1{LۤY#yœNY3%IE_E_5lk.+6ՠiq8:8)p6bQT N-8Jϼ{YJ2HG.B+R/@M$FYt},{ڟ՞sI8sԙ{я&!@mYEXT4ؔ&Tϣ:F->ė2j9, QEC ( BF)s_wsw[Ofx0à A hIжWss>w~$> jdC+Q ;оiZ|bo{d&\~/}ケ~3yDbmtxw\uSvcvt3#5< ?߳K!q+L;#2tE\ej>wq上7(7L`W\)"2F*ՒEPW\$`HgU&Wt*v*SidWoQ\IB=_:p[ࣽSv+vxi鬮9?m(>` ??~7Mٻ޶r$W?}^NӺǽ _/S1+a&,P19&| 9"- Vzc? e"O߾ ? Hṁgt/ᇋg34,>bל 5之YJ{OQj+Pe>؀<_,A'?s&]bH4i4A1k7\\CFtrXG,!!rouArbE2D=1¥s4)3B()zLGypv!z$[s̒9B4B+}w6K> :8'_U-O`1uV+¥I*$т$<5VaF䴳FFC)\Km(K4[mMr= -Uad(EJ\L(zz<'iC{g_j ;2YYz8x14>V(zBo3uH)#{\D'4&r~ZnŴ:9 h!OE8~};g>Ω8j?WklּK\c I!t`HWGcN4w@t-V[I}$L3S4VHORRsB#2r ,rpeEHZ7D4.j8 .*NY]xdcXXՊCY ~< 1Fg G`hJ:@й S8A5n)\ІHo$с>ipT,q@1qfN)97dur ԞRh^hJChe~JŢP}sH$Rg{#`3t 7xx7_ _$hE]Xd p FZ3O$=$g\t6 {R)‚gU- P34 mJZOrXFn*.ԠWqЛ`ql]i䩝Ϯmzw󿃛`C,kiqyoL͐[$Y*bB--^,FWVWZ!6HźTQCC#{y=oɭ)tb(2O5h]#ftt5uğ'Οψψp(DrTv,M>!XX5r1 ۪ȼ^_b]1 ral p奼je+bcѸ7H@qY)-* D =GƉ 72Dg#$0$c5sij"S&r}װ/ZҜ\0zi7f7ˑOB.1dJQ(.FE!&q&9`Li* [Dy[J8 Lyj-7yho8ؕGmo߀'n$>7ه[AhXDh($@iBPƹIebɕ $ <#:j4F{9/SRX4<Ұ>ΜӰq>t[x;ޚo ͎q]0cOH]~zGzq =4pa?Η<٥6UHi.Í,D06Bj8L`sjV$b~'thF J=>XD; D锨ăCx  > nqXfn0]^w9.n3oV ŶE=<|zGzI~$OX;'U渙6&9Ra'I\t$2m'dd5>y*xv}GL }f%NFצaхI4 Uz8)&*'J)XP#, $04!: ζH,n ZfBJ/ؒcd;bs=|ԈH!Ԕ?uZcYRY,.H2"F6x` *kMnby Ha!`a{]rRRlt` Ӝ$jhQxb1J&ijb ˃%%0X%SA'8{&kא}0}%yQz`G|m7e U[Z^GU gb'ӴUguzx;閶և[ES,]sk3uFߡק6uϟ[7ۻwth=Z_9[6زign=7yGK-χo}s۷ y9?ݓ~E=wtܦ _;FSOǟ7]b[nyZ6/K˟Yivȭlj);c}?7R'~hN.R*fuʍF~M@?qT\jS<MBo󏹂I,Ȯ մuW-I9%u9f9> m?<;o3o\O+͓oq=ߏ6=#9&eS2VYeUoޜte)iѷWYW+=-#W{,/swxYլ ^&wI˗)fVŬy<\nЫGV/ CU'UǴaJJJw $Z6_η-aNv7ZGrd֖;ޢ;ZȌ51Bs#:XHպ-dօa?|܎*lv,R^S6 ^=8ΰM؟1ײRƳi2{o{]0FX+cv2>NmOWdTYtkB&?^W,,R!Eq^]Ig0s&$p]6Ɔd`S@wNŧ>o:Zvڜz.hCL$]T7  F x pv5p +H\.9ڕZmEZGBZ: 3/oYq՗r$ӸThm"F؜GZbS{[bubZ! } *PRD(DN5eZs$ғZ)CZcƠS)R6*m2nG)GfA| -4vE^H'wyyN~oYï[leE GI~rI>JDKZ44I1e,<Cl`;8fё$L@ ',F&A`q)o[en=o4㡿jNʭ-WKdKa,$e4MmԊሎ_Ö,X},oqou'˅a4'wr84q:%ea۪ܯE {z]Gk}vy09#ufb~8 /Yؖ:}ӟYow /w4}{w./CT_sQ|^p|IkLιϒȽQ/IQ-=hz6u+,N!|ה|aFe2)/d(VE;x1 m7G_A^ s8[mhs}24):uzi5ۛyLF~/ϽY~7z ^N:b޷h!h:^Wo8HQd܇k{9@pŲ_EӠq8&"4H9w X#cCj2^) HzLcB{?8(=ǣiCN[dƎuyZCg{ۆ]t9X/!RFyFH9X+!\ gQC݋<19ce 5Y#+{È{wZNkp#CWtj9sFAyZƌ 4RZ8Q;8ǖNc3Z`(4wF-zQ9 ",%! j@C&g#Yg Ǔhw&]|?$aqBy'nSTYvey"˒ iv˲TŠn?rm)fcSREfo -Vb 3!dB37!ٙLH~&$2ibl#Hpt2P1UV )n J+9{C5O.~qNR&1prXcT@pbżuZqbrZ@ 0z;@j9KjܸHx Nۘ 6ƈ7qx=ADte,`1X19@2m%$Q -9+R&$=܊Зea ]m~~g]Hupa;RRI$ ڢ\ol3@9 *X0B`D$4M7aS2(R$9Z ]6'ȦlJ>:y"@o uO1,Hx)d$%r2hQ$c& LX/]3Mm$8m5v n8aJ05ajKoMuxQEqVx)C0!B*C˽3(d,``1ͭ,2lL].z8|/o>n{fsߛc]RŪ]{ޢÆY-7@|mz z4}pΟ&r/iӔhm+&˟ Ȁ]Low|R Ez0iC6f[wozj&|zD'B"$hX}*ݐ/ \8J(sc:7I!j?H.2m"\"IPEHfy9޶f3,_No- `A.'Om"K OJy@Y]8 皋RW@bU"KQWZF]]%*yV%+DU"g~K#e(JPi\G]`e!hLi`BqWKHilY'M (HB)ru0JKAeQ~Fg'y4/'va ʴdoL;2w 1+17&ͫ*Z/ff?vJ7yHzGx^}tVt]TkߺB]u պP0vY{ Y z)Oc.*5X@U֕Z{%&|ϭtVm~Kܝ/>0_C$'""}b'%NQKOt0*'Jd'DQ}+ c6ygvY,Ѝn 5@ t?om|j`nǭ#aJyZǪWssr:Kh%m+.cf|/($VSߠ?X)`q~CR|dFB2^,C *$L2T*"CR"jxT~8!,ȍ$S{+%R.sl4hlb[hlFzD5$|qD۷;DSWEX"̵Fy=3Fa!CG22. ƌ#.@SdY5P2zERPG S\-&x#3  j0DPR5c6r5~r]3̺҅PwpK:ݺNRe 'jP`5/}[jT <dE(A>)b-.8 40S ΋e4$x69`^BM$€- ʬӬhkhlq9Nkwv?6.>\HsBdG]r: >0G:Hc2l4Ҩ݅ɤCRǡdֈӈF̡ ;QzHL*z ",iP΁KFbRz^m$#2f{_Y-UT". x.Dqt $ Z\ZCFo5?fTS^u=F:cY.C;H$Nd4 H7(B2QZRaoK1amJ>@KoZ%//Mb#z!Uxd!R&f֔1тFQ4`pHK^ 36߼M٢n S i zM/Ÿԟq~9 LMurY ,Rg2DO"k/%%D )XZYh6䐛w) #H|7>6s2 JтI°"xZ_gPûCi'#%$sdv sò$t\]X:xbX9:n{vVDԃܰo i@d6K2=ot xPLGSuZ܆ tvqf{)ޡlqGRRIU؄/_`FOpGeL ?Q?C|;n }9H 0"}I!O;阗-@g;=89/^c#Hrn[TlI.(e38q|[j6Wd#ԥsx4Pw%ifxMy;欔IMtǡq[<J u6;k_յ| M7-_T'P^9os/]r&HũbJtvfOidVz=c𔔥<Ed*%DQݖv499%&|eKZoՓ23eiSﱪ9+Bν9ڟ)^He|V:ml`m $K UH)u`s&=IߧŎ-ro/|7ֺ 1q"k"ҞR !GK=9N=>\BnǏB%#œq0JBm׵>uˬئ~e+"[0&sE$`"\B)*!FdhtMlt igRi6r%᭶&H]*Ř[tI)Oh,A=X:uW}e+ZQ޻tLn^WL"Z n?LWwm3 :1 |C k$ՌSG !ECH])zB^&0\ncGF+ad(S9V൰EɕωP,3\YEvS|9E PMZF"ƽ0-f.\f')OwsbLS iseEHZ7D նq>]Tb6_ ;{ѓ3O:ف l Rs*Hĥ/f - U-1˱u5<R1*U|ioXљؑ1ɣ-e~#| <4ݰI2A5E'1J@fH +*I:$/P nVzv/0\j9, Phll/ZmKz*atm~*8 ] GC~L<<~Eq%v0X46 ^}n~vvMm_O-m6v<_L8RF73HJߢF~uzPDBrĉ Ar ) wƒd@f.2-SYd'b8k~ nϷ=cz:}!N(@R GV¸UR ^x0Q$0xWͫ=$w<J}D}Qvg&aOCnHi>C6$RćR gDfN*잵r|>%d(A +54rg>x`OfǽܟyǤ("9[oϷyl|t/9G]-Nk0Ww&'@!;NR !6R3(^E07&EH,0&Tq C-ӂ )&*'J)XP\rU0-t~E&%~!kynIN=v'&_?j '2LSEC<}ӌ?KV^*_i@F&lAdt-Ӓ6(Ƙ7POVph Vƽ%'%UFuD -J/U6F)D;MM~y֡DDy "5$`Y^3gǒ-6rѷǧ^g9Vd2kkp-*L%rLav˭Qx[{۵MUqȍ*ٺf"]`룞E3Om'櫛ꙻ :m͟tm75Ow~ WnnW}~tw%n]<,<ᖮ9 o3/|h\ uOQEYȭ.i9L)=\F/ey˾9V{_s lBs !WZWJ+\BFWhb* R*[+Y*[YWY j-Ǐ\te ~Y~goĶ/a: پ'g`<lMcPP2 eRZ/|zk!.F\ L\ LgkJN+LBLꋚC l.ӗWZ]R W'^Ҝ]6yU6Ux Z+\B%+4尫lKlZ Tzp/^^hds]ek ;\za>++#%u4oûs ̆W,mI ՗ytwolȲ4.:Հ ,&2C9 |kJn<';xF2K*m(T*O D geV LzA8KcTZrm9rPwXyj:dA@w ~h,y ]H#$ӎGjqLGdTA/ Ęā3JL&Y `KMxwH`Y0>d^ z 5Co  J0U«r'UʝT*wRNI;r'Ud;;pr'5r'UʝT*wRN^E;lʝT*wRNI;r'/I[UNj=*wRNI;r'UʝTYb*T(õp4W 9~ș=3xpZ8ϳRl#AR+wA+dK60׏Kttmr>|եyִ$) @(F)"P'YV)!qz^ q,tJ`9%R*&X̜ȸ м#0G¥}.^Sw=cO^pux3pVFNQHp ''1^@\Ғ`9H)e rg#9YMdG^3-4 Ob|LT-َnt˾vޱ/jӒ}گz#_8M(r.$΂I;mT(fTo,(`GX"ED I1xX̜x8 畐30}FDV""LED)!R!Q&IE9~-Kg\pQFI{(nkFbPNuCW!VK*w^;Sܓu`STM;SgߥVg\HZ>oσvP }4O!YcٝEpa3}|[cg}~:?5;0E5s}ח|~ AJS ĩv!ڄp`2Qf$ۭrȮWl`ޢ\ ȻV<_w'<:fu9Jd]=t_]ٱS)'Oԗ5]zѫİmF=kH-Mov w7h[Ow| <ѕѬ('[H_+WiMݿCBs-:93kwW.{'w͗Ş3 lR88@Z@; !fxMy;朚IMbyx;&},UZ%Ie_chY}r:XԤ8y M}~@A*NT3|ND%sπ,("̨$1ꢛd˷O.|/W"b7T#5QtP|_V*d>y}k]܅8p5iOp)\&xʌJzGrۣٜg'ݾA$93vlqaܕڮT -uˬئ~e+ [0ٻ6rdW?O'_009/  x؈"9Y?,ytEY& "U>.D1gB\.^Y8$-CF歑̀ :-)в4X*JRe.q%Yq|L@!Z3& )Ԧ85x-zz†#{_ݶ E+*<0^Y퇛lr65[y4"4}s-{E3fHR9gG!QHy8;2JD1DQC.D3WJ"Kkޓl&gFGE* !5ϸ(#USN8e%# y/!VJ}\Z|NS> r>΁RVoo'1;Nۼ{= m%`cWLWjpTz"huG /4x4h)?̠D-gdi QqWBbU{H-%$8j<-y3?o{LJSGJLNL;1EE:VT{d^fo?8'opmή 3hZ.$1Ae?LZq;LJi}^)QGڎs$'o t6^ o 4\K: 훮m;7Jn< [.mC u_m]Nw޷{/A3}/AwrF{44{lh!~kzjqp?,D] z`?魠a+ >4#|lأpRB !hfa Ng7k'Z&ÐY&Ny{K>}F釳~vs)b[IշBq@ԢG#hܦ}:?i cvi0g|_h0vE Sj0X5rkwg7}hu!J 2x8Eg~)>eX1 }oi]a[OQԨQ9?G՟sTQ9usTQ9?GhCz\H+Hz@ZH+:R)ɵ8Z(^mLT)8eo_9OX'@WT¸Q ^ IFEGo/dM!rtS{+(vߜ*4=,Ϟi8.i҈H^`5QK'U 37xYC*/le5Of58I賰=ҳ&Ƶ@]ZnILKI"Vѐ3ʩ^Y'U(AKIn]O;F% 5Hy&JbF daK[;ޗ;:G88|"XϩWR^PXs1^J,` K4gk'9`oaAoݥH/!m!~T 7 ;hFb/mI`^w*y|r|v<~\RBKkR2D'RȣĽ"&i*h%TInb9g{?7I+9Us.+|eaXoix|f%2DF]0aF 4*pEƤ8&d 4!b:D幅bJ@@1ҞEʁw4ƠiZ1y'7P'OnV, q["M]y[9fSxl 8R( 1O1|!OwC|VN |H#TBxj Z2Qe-^Ki p5c,Xoi DI'Z"5J8ir^)K^8 '@-- /uJ.%%b6<QB!e@k"uz-wVz8Xٳxg_BlrF>{atdݾ.-TO_m=8 ܐ~:K,l evǣUxGog3~!-VѦ9,Kܺ ?:?B1mO n[7]Mw[6زg6yC]xR!W{}gΏG{үgP??Rt;+8>r-ZJ$8Jh &ƺM|s ,ՉŻWS|0jGP1|'gyHDHKJOA\Vy7lxk5mRek%m|h"m$p>70tW#S`:ߠ[GA٨ Ү]?.L<%Nvγa ArB?~djD3GHdW@:E7)#QmdTDpŕ&^hA/XB$4+& Ix͸DN*2z g64y|LyF\ǮePx_~6NH2;eJP*N^8WoGS,a xDQ*$ZP#'ZU2*!FXeLNra8z]@8TNEwJTiXLݖq;J9-,63d B B[mQm.oo\_az? }5hp8:9AJa ɔ0IA~r"@ɐR$K9I.֖Ųp8 q&W"jf#Jh9AbT騕R~(aeb(~oV+FH.G62"$yj!.}!~Ws '! {cP>]B,jEHϴX.f"ʂspmXkNHN8D>&SZS@U( KDM$*[gO7wݷ9M0Ocsʺ!vCo?޼FWM6]6i?FO: 3$@<([y\I@BpZXDRy!I)m-N&IŐb T#7B)|Cwg #~xKp$ QU|T|ُ|)Maޫ&NH)#2r"ZsB{(~s޹٢y_P h،SLphCjr H2Tqg`4n%vxj>6}&³U\VQ^\'2x~<t9K30җ,l~Ϸhr٘;p(n?ӧ05;ƨwtKv?aZs\i!6Iȃ>^)Qh[=kqno\B~~B2P-ɵоjl bX~:὇#@vz3x]jo; j~wofwѴ3o_>v*&.hqK^ p6w˓Av^kmWjpߗ;<~t *LjII#'7Q&jy H%j=3gS1*x  /@RN>FBvVZOj81±^BMk\^sֳ.OZ-mbf}P>gGj=qpGL8sT/3peAGT,j̀ :-)в[mR xgU*ku,` ՚1wTHT٭kq \68 <ܻXՔ3Z)GyE;Fdi#nv3|vCa0u_:k|WHAh!:j/O9dR %MnpP(9}+mcEJnyH72Wǰ"9,sX*ْc-l(֩";gY)?JgGgP|f}6 W77Ni_X&Q^=_޿1ُwI,8 Q5"V)YCpŕ/pgϮWbǡo ܧ lN$qQGiRqzK3gW p%*\;3d TL*\Bg >M`E{WY\WYC <PJIiWW+su9\]=M\WYC>Z)MWWDi#B5*ˡ/p*K)uWDRg͏L\rTJs5NMSS:5NMSS:5NMSS:5΋O#B Re Wvlmf7uR'o(\rqJ4U@*8sNgBP!M`8I#D ԕPu YI9ah֕"x-eQr$k"$LZiG*)*kqq,PU4zo*L@v|NSG9gnpyV b-_ެ8\§ᅩ44vuTz"j\]*WrwFh^%e|#,`\YRc bmOGX>dg/:zt'+Oո'-٘MN %Q( 9:@PrI ƧT \kJ=!&ji߼3&H|2 Y]aVL@zmՋ`C-\.7X$MY:,\ju7lT¥:q+D1!^~7| K"P4 #oL4+$ePRugZkby R*" @0Zcrȣlt.D -jlS1HL451xC "e`T<j[xR9& ۊk_[h62Y]5)${|lr7.F^W]њYdad>3-]w]ofm־BnmgYf"뵞E1O7b[ϟ[7֛9nfm~mزM7[wv;vycywQ>Wg;{wx3řy#)G)$w7HSv0 J bHi.õ,D06Bj8L`s*ʆ^Fډԇh%(at\`RxFLqMD$4BhS8όY;^@\{w5{}|-"kכԽe+;|z'zJ MS0$"%t`zm]hxޘI#=O\Pd5Y&KL 7)T)%X:iFL~>Gt>*D l[0̗Hp$(H9wMTm u6*|YՉtޛ >WSmL]I,oiiIQW#?tw{#C./!FL'Z3(R5k4P>QeM/њ4I1H|>'M܈Þ @crO*'cfJ Cpɥ"!O #F:*qQrP"i  2O}F1.P'*N)}|پDu!-4NzƩڅv}y“Ƿ?w9O k3ϞV.{ʿ(*oOؔpܽ-:1o >@(F)"P\DN5(2JR9}p!qzυTHƱ`c))R6!HFblG,$XXlf셅悐Xh* WZξds;~/ޠ )0A $brL("$Xi"hcrYy| Y,&#/IN'1@YL>&|gL)q#’ry.]lvڴ$jWQ0XO(!焓I*\4$ӈy 8!.9~WNC|o4IYYPSe9iI+Ò1Z d`Ev< %p) bT;TZÁD( VHu mm"ZL&f2ݔ)W隣.mN>Oh9 ;daLѦ}=up?xҙ2SdD gy༷<g±$f[^ r7+0T`80V( K*1e(7. w&JTB9b!0TRWq[_73CenMb{ڠ S")Vn"#yJ[,OfCS!YW&V>3 d%)c*dZ'_dn00*N_L\R9)LSńxas:jp[ y$DgU_l_XWŚ5+(A2&ԓHQim!  p.B38{0ΑEIDf 1Qit2$ Q}<+qFQx)pF|1Cb*X=lqzsɁ-4_FWbO\ګ 6HmKF-}9ܓ.&[jEk^Pjy -ޮ7O=B`hoтpt7lx)KHr_~\i~^ }5m䉔2"71'ѡkLV[%dډD̈́RRVX'N\ss>,d+p@ s]SS+YشNT2+=1 xjCQD2'ޒ"X2gO4ξܜKvo>6u&³Ǫs.-%끢~\'DLsٯ/yCR 5x_я'!@v!/iGecro~Of%᫸?'ߟ#$.oc\׿_&yJ}yJy V2I8J2!]g햽"]kx\Cm~"EB 횮6o;`n2۶ZgNԛqy2C@J W(7pe/p*K ++-P=+X.;\eq) \ei%;wR*V•\ٻ޶r%W`f0}.U\Sw.0;p([Ohlk7mI 8E!Y"klp`RupzJ8=Cj/5SGp\OZy$vpbp+աSX8Zvą":y*R.\ %;'vU6x6pErž:\);v&J*gW$c q=*gWEʥ\Y@vw~3y ~[_ :r xңs8zOJYRƠ6x/`] UX~ZrEoЧK>tcG$WO? 'Ӧ&Ok3a#F| W !>֊3@,9ss.\Nޒ+RbgɽEK53\ap>zE\WEZWEJ\ARZZcWE\</"WEJy!EZif|"WEZW$a7WF[/[ YUHiXWoxV lH'o )q{/"\TE oߖoڻruhm|UDMK@wL|?SkX$GkQȂy)D AeB0#==e2,߾߆ LsM?Sp6tm|mxi_T?_$g7q kI(5Uu?nc?6\;~w$\ο) W7WqoGi۞~B# l\Nx"ɧoӬKY}B^'\7llFrF#o%|@zCzC儖+ͽGiw9>zW͠rʅO9w :J'\OԪյX9kȬxIjn 灁'd61x>iSpX/dEuLyf]v]{`ԫn2-axӏ-=ػLބ["1ybz?Mܢ }auS?ijE*qW G֑>-g<_l?{'ȭ6*g]],y[UvFۦ]v)f7{݃iK^,k۱,ЧuXh^B%`w_SSAeIc?=znttIzV/##*n:uy*xYRRanS18-H\dՙY&+ڛ")yZ3`H ~ۨԒEDQE"nZO֫pƔ&z6+%}g|I%ŵL&h4PrkUFjlGEN鬳*9Ue\.vxҐu8x3("au_IҢ󥳸X=K֫o/νH}~}ïf)ƈ۪C;G.-UFp-|(q,#NUC;]aaGcpF?RD ;ဌ(Aϐ"[k G,= J4̐ƓZ Be."ErF&n"S2-j8.4ףD{L_FfטcZדy0W|xt R؞sm!6$d)I$ t+qI-q5K]^dJ:*W13`m"QbH&/'sSk;ҧ}12x5[H3繧ھf.:6q7X|eb _4'/l{%KtQuP~9Vօom剜 !Tbd!F'W+a@B$ǵ*V;Z5qkJ x4NV\564cz,eY G6#WO\a[m|?iHᴪ=SoQAX^\'ڭ$fr+x, p 7o|ҾQə&NONN$N`%˥`JpJ 2̥)/G; ADRR{=umvFnJ]2:9c2BbtGpBLNY,E%s,%~S0 yzn.=}o7s@YM/πkXݰO`EILg^ǼB0V'eqcA ,Vj 70$k\ =]ҒŮ`3'?Y}&#l)wF#1Dq.19I;1EFOww/D[p?S?,{iЛP=Ժ9uS?"?WUb Y KJQP0PWA;/4oرͻ%a÷,QcL1N'u).E"5a37|oq`fsgumJ 6Y?-f[} ?((SH"HրFrJ(Lkz!&-UJԀ \p]oqGUR \ %\}֗][~| 12+0|? ˻t~hdNʎL ? ( Ӈ_7YM#t Ur(~FiH*GpU:2us \YnL'vUv7p5 7"nzgϮj\IrF\q'Wg̎tjojb+{W$H.A_U8JmvzyݯBGpI:i?jZx"1(>/rO+7ޢX.2U)>?Lm+n/ԅlLU]V$`w+[e{Jw֒FD&Bpߪ$S}-;ls@/&PϛۻF̷} ?9@ƤFr{Ƙ5/5_wٷ\M~xnޗTn^YoKo]4؍SPtҪm !yVYg\Yzf4nY9zڠDb*Y#D,AYūjC{YhK9{+C[+w*d hAYhBʾ(S"%apA]%81)}6);Yb9>FWaKby4>LGL7o3lzоMЧ~-:fri ,K1) ~M m#VT1)fz@ Bˋn< 'ң(Tڢ, J d@1VS6:dc[U))ę:0 A_'3n.6lӭG.@9Za/5["mEVSЬMTV"E!p1MVJ[++Dzgxz=6q 30G9GKUQe1T|9(-~J T(N E-5`1U褧9*Xpd',X@١èo6~`]qdhWPO"EU&A6K-@A?X|0eE' e &d(`R 4Rfht9h%Ynj7x+F4#{s"p{k}x<Ō trM~1A&~tGe_}] Y#(*G)ﻓ1i8c̻ͮɢ`u%#}S<%+tт P~Z-las_Ilxȷ\Tcטk}&ٌRh0;o1d%1s?& @+jG&ہii)kQ4vR4/15ZtnS0ƼAFO2Nyf+(|3lxM3uT=S4l('\prN)HdHl$Y]H{e7m576V8$ssFN.*Q'J @m`A 7m},2KJ (3o2f$NG͍??_#%W2,<PS (hs c6:"dd2^$o(`d"jeI Ҫ&h T8E#$-iXȔ`>uZzǡ-[jEA*2'$8RZĤȬ[F (wd511 3ljrG58Zh|c墚>XGl0i_u,(h`#bS%[4*Yȡr$oQVFX L&) >k%TBAddV)F{T%֫$=Y`oneC{_nM1WuŠ,O,r')С m;ċiib=fґw؅| _]ZAX>Dž_tsA?4O?7bQ y0d]xחP" u铹hd]P:`U6j*ܱz_%!g1&'L :PyL"NHHY]Ґ٦=$nhyNZq8]}wmlCnsI:]4c=fꝈi$q x깗??{Ǒ] O qE@E=mBIk(??sgJb9a$=n{NW*c6ڧ~1G ?6^^fe<10}:Yb}|3:ט&K(U;)xvXg>&M֒c: ΑM&bZPdZ AK&[E',vXM\rL[,nۑֻQďyyuM_m7Kː[V2j"A/{SO5!elD-cj*IWˎ\:)j)z.[euh EN@0Qu‡ "$ i>#ٜ\~azX8w2z{W]g5} K_=o-Vz^vAڊ}t|k뛸z0_' W6W|{4ꖛ^^sZ9>*7r•75'w]|ͷp+Wޜ.~{w:Mṽ-7nZJ?nSᷤ_\z7yQ/OQVnG[k{4[{3[ڗjS(=l3d *]0BV=yB\}Q;ݭE~MzS|v⛃_Y@p,0+ @# :K>~ KYS]9BNg?iu~[m|&+N~Q.9|֩,Wv_I6ُ>xnv!}/hlONm֮E6OsA*^uK緀qMU{b6-Pd aǦç8e_7뇺iJyKg>5qXO]H~c%l(XG۲K"oh UgKUeeGC߁Q[1|O_Y7#,g÷D=7URvִ/Ng ]bs t!JH|vحj&ԹZrSmr)f NۛTwm klC:oQh)EL$dK^-6ڔ,}n"ӷoߟ<34sc}Ni53\G&z亦ܕ)aL",=~\-y t0kd38p'-Ռ NUF-9+ZpKFCC(Jd>7{ydڦb❘66<ʒ&ӐdKD0 #.|KXjGGcT91F<ٻF>8=bT:䑅niuqI%'h\bjS NY^dc)_C4*ܽOw4e7-GZΩRRnm5ljxIl1p[c;S\8hh3!%UTJJ@j~Scuߺq)hK -FbXҠ,EtHH(k 6Zr$FbʰI&P}j>VqC i<Ό>tKwr6V,KT7fcŘQ wPB\3s.?ow -;V11򃳣ulmFL@-$t>gu`yP\K_!&*fFVX94HQZ}OxHPBhb2(kvk+7< o 1t`¢Ltաk4=!W1dpiۡ&[]+!IxL$hÍo~:Y.Ru+`+V  XF;KrIn hC*j]J3n@׀R"ðl`E+ 5h Dؚbؐ:#t T(VWC+`sܢe+oVQAѠ;#$l#ca>XTY$@fZ&הA?Am j @!o<yg4\oF,Ұ*+ǂt]# PB rf:PF.+`7,3*@55eV3o= ZO5Ձ{%AE joQdF -m5xbȵ6/o/Wf5]KL@@]@P77kG֞5`M?3;Z,zO:5Ws9&mFH9[6nP3Z|7\|Eia 좃U0mE:ЁĎ[Ti3 )nA-[X ;Fc_P|_y?+AZ964A'w #P}CE7t PòBc$itrQ,FR<z+Ϣh\SFSh pN\76xk+FnQ4xAc-Z Rת&_6LzЙ d *WZOk[r-F_Tm G|5ěk >FŢUq`h`EH֔7rLAZ #C+3{y;Jwp PH՞f((qK^RZG7$ì -Pi>bt~ќ-jUeuiL"r9RQ Y 6ɻW\ER!6C$](Z4&TVH]ON ^QD'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N q=_'PN}r! pc'~G!J @ B6H@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 |@Pvwc^KoQq'Sn'M#.䃓 8zz:=ߣ7/~4mYNQaӏavł>X;? 7O5ƅGwOMΎNj;(]!m}=*}X~RyoVoqv;~Rrw~Zp2 DV)GSѠ%^Nf#D'A/{ "5JHGB3C~P/H/Crrٳ*& $,1 IaC^p#0Mfo)YQ-fgh}ƉYXb,,fa1 YXb,,fa1 YXb,,fa1 YXb,,fa1 YXb,,fa1 YXb,l„M;Ǫ.no/VYعJJ2* Vur ;Z;VE>zΑ?P޶{d‽8VUH1Z9JVbh%F+1ZJVbh%F+1ZJVbh%F+1ZJVbh%F+1ZJVbh%F+1ZJVbh|V.Gɸui}08x /5^/Wo ǔz^y|ތ4e5)ͻua7e<, +à z#(wRgo-{qŗً Vh{e]t"jcSRüQ6^Kz Z^k$t<" QLʵY>6S.,=Md׻=:s"z",FQ(Y),2 Ekچ>{z0,=Wvjͅ6. |#-'=?OW *uDԦ0KQ"%B3;˜'2͐WƯJgZYa~g]r@_|ԭTW>-we7LՂިU_kimyq(K'Юjv4gPR^*嬎Kt2Ϻ95 ]oY]a*f:u5kכ0.!}EmT٢o 3rG`>_v:˿NP~d0uv0|,Z!dƿb?d-~֋K7zZ ^.Mmm{/%]{:ޥ*28 ew9t87mA/orcЯ;f_7Swò~Uʅqk{B njP|g8_x$LREev"KPuwfYkXևDga6_D-d+pgE;˷~rIh{+Y#e 6͹W uu,_2J#U!8> , )SxuHqfgD i- ‚8zj3L_EO_]JVg}eأX΋8r+7u?=,Mwqjy)4,5·oPTf;a6ȨH1BE# tRA+~F OǾH2#"9#7i V W#\:9rE6@dND*ƭIsdE0eDllGǛCmˢ\̦%Ȍg\ܤ`Yj2xsFI)נ{Vyki7Y}zި3.OiǾx(3{?Guyz5r7iиY:'n#_X}Ҋ-qOъ: z!Fя(NG?y;g5Y1|YtG؂q .j,)A3t/F;[~a}?3( KHâ!A):#TjVKD4;y9h5dg ϑ5.ZN&sb=m1%xZ '}J<9Դ^/k~)~}ﯸ sb t DLP-)f%=#Q!rg`83p"1]e$RcmdxJAP$/I>ghH09A.DK9 ޥVY:ǒ3GMKPM5J`fTMYU%6Wf{Dh>Q iV$#Kصb(^sy.w:i.x8<& ;~pKP'y5Xy -8Z(/^^NxSIk,O`qlفbR_r6E1ݮ8[odYJnYVP p{+ǣn56{~sT(+0q`i )=G2!LH~B&$? 4E1&#Hpt2X1]}SV xcޓ:=V`q؝if ha[0ՄaFhNP7NK]U|JG9wSbОכ G{a]%~w\il6eS_Hb|Ɗ l dAeNF]8f`T6J!,V ơuARL,v ݇EG_ܲT:ؒo90&0(`8 B9N^*G(;v1ݘ8*Cz)9HE%K38;P@ 8E?nN"+Z5 4݈Tpb6*ʃ]ń-gvVHA3vԜsJ/i~;~ENF.Q_۳6ЊB2 `K}E3'^?INӰS3-uc:®)46i|J& 0h G>M𐴧ɥZ58;XE]ߛI0=? n?'pb~%J;͍{x.%Tdj.3T_E@ęwUMv/i1'{> ; K};EI{Wmgk`*|m73_j=c?t?ބAg !+ՙNgoW/چnA!$k<+z4tLf̢1Ŵ;%GvfP3Zsjoda]rl\J.qԢuRM55i5Вv(u:8)`Lx ^0oNnj "c1A@GV5ɑⒾhgG虗̭b=@F*jFTTXi8sST`H PtQtbniHXG"\JlB OcY r-a^jF>ͨLOh;$IB׼AzퟨZn/ iwk4ehY;]A i A6jW6*Jt`y (Ee *^LɁt?)r>w<I_{_:Ɲf>+tpwVI1*k|+\MnUܘѼ ƧJ !u%!pNz}_{(b;+di]b@rN!\ޣk)(] wtY.C_ͻJ`hVtb !zq !r&~OMڧEaP`,>ǡ71 Cso* JTnطQA,׿t;M~,.QYF 7|ѧ9CR{ҝxr1?+WN7̓00Lτ3{?{WƱJ/nK/#;qp=Aܼj8Imy0}NuHl`f7k=u;r]뜒({oƣB_~7]{Iڞ_SR G' 8>Ln~ZN3Z19ymd]seJz&s@Gx~8'8{kTsQ4gXW>\ `E_~~_?^_xuq< q~PuCx!/F-Ĝ2bi)}y/;CPV @2Exi^8u櫯SLgI˗#̷X{ Mw2vk1MzBESJfﺾ`!+V|YhN'*Y%5"%Jh.¸{%c0Jgh41X'P/ezGdvVȧ Q)*'5VQ?0ëcH14bꩴ}t!a8g`8 xTӘTT8@oH ="EuɖI QEA Q1^J'G΀H!QKk2#BVkzꨒ,7FLNqw&ܯ`0V 3M9]iim*Dh6WH8-%& d[6VXcy9T Ej<[M-BnDXA{(l[x\:oLl;eγ\E~f/wb?$%u.ü!m1p !FZ$hw1xdZ1d*jo#wޞH~l TMWWsfm,؅%o$0)}2(Q07װ. Ӳj@ȉR4gKrY[Rʹ⮥Ŏ լYRSJg 8o*DHJ-1"aB.\ (Pʍ"S)N"1rئ|6 0)$i`p\ ʉO }$s_=Pұ` }}D u0_J[e>\qXt 7_L<`s-:٢ǟoFҕ ZGK!zƌƁZpcĜE NVFߦ+2]I-=%I_串%^R>IZacI2Yy7| $fy- ˺Pɿ:yjXH 6Nsu (Fʉ'ǵաPGZJ鼰TGBƘQMK P.Zw2$*Q8&܊r<ݷj?lK;"h! 7_;ǧ_ MKi$#ڨcLANm<ȕC EtT~cZtj0 .˷W7itΚDz5'Ժy^˳7>)AsoAjlΛ|7Uz)aeyYf"}JW7q2TO\\_V>oB> wH"5]OxdwAȀc (Wۜ,CXM,meVvoe#ݽ!{p1֧!RFyF(6a`b\ g # ؅QG Fc"RSFD1h#2&"ݰp6ָfrw@ps :&?b=Whv -xs3]}"[-?r S 1M-cFIDGx)VFrph5Z`(%h2:ڀmTLT;A i76'x)z8.٣=_Jr\.Sd\J,6 {`r)svLr'%O1YQS0rA%%^Sth:-泬zhp݀o ǃpR{Z~}hh8F\F Dad 0ggN?'GG:#<[a'.*;-E.?=x6 yq`6mQEX[V6~6y[TuPmT4+c;7,3 ;kV2qɷ䵜wӐl'PcvCM:hʄ׳b~㖢B"e̐3_}h"Z(00kLΙ6ɻS`Q&6Z:h!d"Ũ]^g7Gikl9ō!APig9ĻEg;>9lY`x+:64y7:E0ނ|]eUJSh,QW*a,K!|K n#d#=8Υ`z%׊6oE~ `fIiN m6ipf8m3N mӃZ9OIZ$NmFeD1jo>ɀvq=t?nD3_n\ӾA%lƜ\џ7p d|A; %TԢ&!BlhSIH eJmM8\4 Oa &]8:nǃƷSgQ> vddyQou|٪1r>ܪQQac {CbZ#2sKC'$\.%6!`BL¥B1kQR%zqyְlh{&u4j-j8ߓǛ駗wBoPߟlBf Mr3 J0IkQ󈸲Q "/^y (Ee *8$h>aU,)TD 3Ќ;$ND<G1o=A0%:Їu16HVJ{Z G>`vRgђAxs-j YD Vpu=" OZ:I*I҃[iͤHfR];IFUNrbR 7 kZ񒇢Lp!p,zgqb}HDg/a]L)j:π)l'|:VgcDRupRnβ?}]3՛)'b\Bqڗ ػ0 Ô`@Z"ܠRP;1aۓ`$\|[  ʊN[lBNTap+U1\ q\4Box7]MR,Xυ CnLH7m{^o>,}Gs6xھ7-NQ+> 3D]0Yz= UŃy`v0LO5_:xyU+ճuSe~xTčM:3zYOݺnHm7OhP. V0bj>;h8浽juΕ)5u[Bu>RÐcX)>Yt׽F5W+@uzunVtWŋ˿~W//^XQa<[pa5-l!tܔKKˋd=woR2Y$xi>3$ԙ);MX,aŲ[_l#.঺bH49 l/MNߟ+(.bS>R Ϥ _%,+K%Jhd,dB>C;"ЈJ@*E<9=McRwSa㜶BҾ wɋ4MfAr?ШC"lU* 2E Ne(5ʈyR:)8rLZZ˔1zZ:0lIߪhh& hc5ü[ngmL;GNjW;N.KpU~v*HǥKhp ~\-^?.Ӷ~wX?!o'u;?~ɣCj$CX:"tDAGZZ O"Z^@вZM Z9eSجTRI2fJ% yQڦHIP|T$ S+(u)2HgY+P]UiMGaHlIލyf>!Mކnc﫳+0pBo=9! v0;ԍQ7;q,\w>9-Jxe(=Ad۳Qs:BMt|hhmN*!z' .aʼKE%NHM9kTI}I=KzqP( UmzC~wjFz[~*EEݶlzS#Q=M~/-$2SՋU cCr>\i[QUK|!B]??]n&[OQ<u#/j+w 31{vgusӉXNՃueulu:tQud騊"Hdzz:teu ա(Ĥr%唜6Ԁ|aP90&@2)l ^Ma4QP::~%gĸ/mR_-BP3|HJMsMr]| Y7ϦKf_ـUh\R*t[|*]%E@벜Nŷ|ڰ]`C2x_Ss(0$BzYTY18IXiRlmmlk+0`trͥvWf3#1-Ǫ">O]Egvq o XO&a_9;IvT*ԗ~eYNƾ7]'bQIyuBșr Mp%#lLEm)1_" M{}}J?Oz}c'U͹&s(R&Jes%qRҩVڈM'MjFaTR (.f*V$o){2f H (bu1 8jMN類<1E5NQEf(4:Jד^_d7E'$L—“7V텷'W1 S H$Ƞ!LR"H"TBmW[ JCO>T\sũV^JB_Z#c3q#c; ͌hƒbBڶʌD(.N}'`g7t~02O^9bGh +Yx HS ,dr @1k 31ri/.Uer-UVyX Tdv5ɀK,Sy tSJZ0-fǮ-[=jEA8(2'^NbJ^ĤȬ*E>+%LM6LYCffH樲aIYD lLs#|HTg}cs~n!b `<M +˂E9k1z$]\Lc TQ1D. V\dYN98n>"3ZGdP@bG BiC Gғ(LLrL] |RZY d)N/Ԥ0#YWkocfjReqkV;6?f^iD|eH-beDxQ u +#/Qge[t<{hrzY ==6 30'A䜣qLjSS)Ei[[bl%jk99lI~7[ _XWŚ5Tт8T2Mg֚u. XB6K-@.MM's Qdj!Z&iuЩ\fQH@l`2u8g:%4w1P4N7yz2~4G.: 6H40tm8{һ p;mg,<^Ε}M>/-x6M,@Ewc#=wڒ%:Rhz~+VZeY7mR])*k2()74@EY F5Ŝ< OҹQN\qNۚvM+5lج-RQPwhqFZ<,zRνr S m67vMjN$Uk[^xXuƱc0=NH۝QHmع eq0R[=t!UJ7Rk$ 3*gWU\'XZ)ܩUDWo4N\UWUZ} Kpʁ0X`+'ROzz-p%ӆ^wGOL_@{WpLiĮ)R S+)W{{e%Ew} Uטs*3WSsMIx0AR6ÿ|f-*oLT\H0]:"Ut&8yT>g77O~ ޚJTyC!gғ4:F WDp>g D+ / NVM[ F ^q;&DiBXДp>]ѸGjm@Zes*wk-n#o \O?\Yȃox]p"oZ Q}{jԒQ0Y҄-C-~?[kM*Q;٘B[<<{og+ڻ P/?H$PC6i[v>y&PRd[Z.P 'w|y=FMnc㶀EL2'ЧЅlB}bB}@ijeߥ|YgSsHЄTJNɚv*c=d͊|JxpIZ[_d2$IEiEPT)(#9iUZQk%S2MqRn! OwKY[9 -մ!/d)jɝO)Y8U3X!1}ņ VfbZˏ32M$lLSU\IbiJiz[4M) `z}'қtdFS4PK缵 VFe$J#J.N+{f`*dƚ.U*({RYH|H`C2.%n+kCN,' o18IXiRlhY6 =@z*!U 7pB9P :KkSiULms앹#)s|v7D_y,c̓krjuBp|3w .Gꝥ\-"2}ԗ+nٱ-/X#v~rJ'd>sq-5+ )Bi*Rj: _RDȒX*fޏ#,FJ˒FdN +;"^,U?^`w1.< *:gh-x>&43(OގeOgͥ͞(">,;;gkV@*1* R\!e_T2NUOZ NGު,m: }ٴuM!طՑ[5~~{ckKݣysI&ݕi[䤳RmX|;M PIz%~3U#~R Տ0rF7> XJ0i+/?mqQ/?(sw0'3x&+LexoP?o>OO }ּٺ)r*_"FL] Kwѥ@K%OyzwyZcU=:g(?{Fr-C!`?8M8ĻE"icW=3HCȦD`[gկTԌhpkVȣQy o14duF:(0!S&aXj1kQRk R5,{vt4,|9,JZay7QK5kOW=O6ٲ,+5k|Uvʀ6(\$!F#Fez'k<\h5ԫ_uxԹ ]w*#R"%1f:wZI%pD AP8@DsNJ0l>=8.fՏͬ]ȟkkPX$#Uj00 A}LQ(,2q"G!yښ0E l^G!@G'x-$J>A҃]iͤHxbK[dT|IN#@a[#^%I)o ݓv,66_k܇dH:?/fſꟅ7y aH}ggfomey>K͜MqiVC(q*51sF{b+k^QdL&n dE#LQbzf: Wv7^@ͻJ`PVtb R%I\uyw5w° xʓ0CZM$Uj~oZMQ ,\2ܼ;⨧ujXK>*\Mm2tQOzoe{Uyp, <Ӌ\/a_Jͼ.hzV]-&v?2Z# m)׵t im-`X `,]|4|OUapMT Zm+Si: sH s ?&Ւi(FHq{džN uK |ӫo^/z70Q.xͫ|30Pzm]RՅw݉m2UԜ[q5-^_y9?.Dҹ ցB:q.?U}?-\Si=΄ujڛ7M4F"2.9n+!~w+Qm x#Au[]oo W7Z F,}$}7$E⍡_0Z.zY$@4&s70IIzhk]ʊ}~?"E 3=hnuH QEA Q1^J'G΀!QKk2#BPgCi]$VGĊu63,9@ݙ|lףJf[Nk䪓38R!>E9[4"g4`[*$T@8(-% )4&ϓOSajA8@"齠 EϹ"aG +h"6%&1Ho$h3rjݷ`{x)v%<~DcdzD= `^@e!̠x sZD)Z2li%%YOZZ>GAՆ@0ZSJTE@1p#B(v8gYL$ =tfSIe`k.3f4j|sWL#,8ZI)ꞡT[)Umվ'=@(dB[_{j\y,\i[6kuOczq:,QX0der{^\,kszh?o^PLF77cr #|%9ӳ_ yXH2ȯl% ֗L;ZP)O'0Qe8?2DG*)t^Xk!cLQMK "l{dsIdb.W'ㆬ2zhO?|B2UVȺ 'DkE5c.L\G>e߿~ܹo^Rz٬I'vPLh7dҝ+y-׽Aepi~l0u]f#x$/凹}= }}5KLsiM rQڰ4 mA%E(0a'm@ڰl jr/mh&!yuOUk_^yߡŖ,\[Tu&fTݫgUv6:wrcצ#!TX˱IzSԴ݀?.&/ JP*]֥7W9ntQfK־n$-7np8Ҩ7Y|]rCxpNDٜؠC(.kn'S9DlI^aś?Rp442)! tFmK?'iE{m@,+;EV&׽yPքz %/5Ө K*0.{Vbz[^.mqJztϼ:?YFn<;M'{b9*&eD3T‹P;HFb̫Uh@Wů~8K,%djf64+rLq1-ǜ|cc )a}\`I`&?>@DƝ!]Ps"V⼸lb^Ζ6\ptp^ H-*mm7A8yEr(Ty]tc 8[vK ű-䰸NW "YP.w-n*uʌ;RZ=1%󠻜Y 3H(򥖜N)bxЄ^q[Ψ68ab1X9qi񺴹q|߼q֩N,!yh~X>ج d Uq؅&(qٖ]Es<$|BRW&!X*8]'k,e% mI7s>7-K$s<_~I`RGTRJA0ZcbaL-m/Y[. I.Kju@r-(6bsOK3U /s"T*@o6&{-ǚ$[qrs[ {^o.!gFg)MRFXh5g{ |DΞu0`R;IqEA?WU:ð~˯zQ*2E K|W/Hh RL&h7C9"N+>]j} #&H,8`1-`ѫ9\/E:n^Pp*rQ0#u k49W Oz8oiK{%:= |q -`$X }9celdN]oACͷJ*.B5Ol.Ol`~a6RӅo,`QR;Ӂ.sJH|xԌz 1RP0vY4a#A29HysQ byb3VplK6ڕ\<,Nן@\-y-އaDqr> tfOۋ"Nj1̠\OD_:+K[eie SPF #ϑ:02#},advK02qjBB!'+#IPNqJ4icSadZ I<"N<f;2jSx_2(T$YKŽ1Xm !e!RsO2j`HjXD|0a&rб{+%R.3fvF 'U/v,=UU]g-x:Y1IuRҮGWV#AT=3xa6ȨH1BE# b*or.wߞC3Uw|1W'[*6w;X ՞AiGUTC`Mm7;+R[.T R>Yz^9 ͈1 J4QetDnÊFG5`ţqYapJI@4!er4v6j|kD >`A*8C[2[A[JL+'6"'^sClȸWx;g25@M1Qa0 И*UF3(0Chf^@/AJXcm@y y W9E)x-\Ř=p6vڨ@@ѵ'Ti>Si4KJٜ+ *a-ch]oS,c9f;S%x5^+k)L@yM>j2GbP󨢚+&9NxQ <^I^8:8V?dX]UaErռ c_uWz|C*NhD9 cseK8[+MR!.I<\[|oE':m@)3w%=woUe^mu0ve?9} */Jz/]îvϮ+dWJ P _A}1*^tUR]}JkyA O@M./]%h%9tvT2+c$>`P ׻u J+$sp^0coaR ki}ov>܇wGYi`6I'C&\Un70Wi܍ӭ7߉ ;gTi?72 w`sNDN~"@LIP9R7@HLnN ?GUQ- JТM>ZG8{Kt+d 'Vr3er)VP&e_￝gg`~SR2o} `W!Ã%瘢av>El6GjMQZEhG5J2ۈʠçD9އ.=`YT{=8h|QQPDKTÏp+m>1. pZBr;$Gɷl{pRtypyI9a'υd;. BMc ^i6Qxr!}uQAlh׼s@Ue $Z(&k~ƋP2]#ࣧGMLj9+j>>z.sLRL岴.}*LI[.٥ :>]n6vZ db/Y6iXg4붾ܛ߻ާnM MzћϤ'(.~t0|}3)qƞc-QwtoqWGn.0m^I!z7H9NyS:'iGsʅr m}sd(D0 ya>x1QMK \{P#i` VFC(Pʍ"VFuY-[))$L> 8t>á) oΙV`w|C D4[ h]JG+b>ɤ@/8ы9p/8Ao7_1h叼A&ai$#A1r&hAt]Y`0ڵB/37pM=ZQ *ϤW욮Sx{wâO.ls@XLWVcT%FRl`JծBk;]yҸZZƇu&pH?CUk:'|Q:5ٴQs4Q&ECٴPaDuwsZsS$)ZS3)Ey|D_A 5/K;༒&wۜ1șL95}D|{c oCyQo\,!mS^F$U%v.b {sʰq/y(rS0 .//M8߼4be]'Fq0S~mN)l&O lG|2tY_8M'i췟gWa7CVٴR N=fYpf;7Qd$'v(8Қ`:RLݱ ­y8 WF?@0\! H ˁ* %1֭. p"aQyc?>3B KuxLHŶ~+8dQ#JY}r6hÍ^GNJ(ZapN\QQe\#O:+'#Z~-߼)^p?O f3=|tM\W':e$ʺ"Po7{Wt$Ɨt[7  M#{ X? ULx{0ZdmgbrQ Z=huΕ)5u.1ˆ>_K]B֋Yb'g?/tZѩ-_gsw ޼Np﷿ L?/޿No`:;,.FRq|<~8ޥhN ]ELx$T:7_g;,‚x;njho44UlиUԬ+4`\,r͸W 6A~wL}t3z٠ :p&y`Dɴ~R?$M$Sr%)o z*4ka"]γH0 HiLn* aVIzjkÈy@:[~ Ϸ!*D"'2Ge ,)-"Ά UQU/_dF%+&&!&oDcb!%|{#>]Zjy7| U^p"+TJQj]@G!ZfBr:xDMU\oވ>Lw7n|Ȏ Pv eqmoNH x̫4{Ϥ}|.ӊ@jVxP 378XC'N@jjWjV/ Q͖;+xø8h&DΝL*P HaDPNv T*WDBoΆ!- i-mf9ڣyKˆ˓둠'J}ND}4;{3G|[*r nSN 4g&k4`Cjr|iSLd:W,}]BnjH9iz"\0@ ji.j_EL !2댫3tvomFYϮ}Dd[։V6fyoUfCevrt#m"yȪ]v.yF-s<ЖKpKnUA0b1oA Ƈ~f^o$gQ!\ (gL׏Y]Ks5)/{B81$y&f *\qZcrS)Q ,R# 24+DJpBsAbq(jӒ}xKEz#G1P'rJc&(r.Y;+ % PyNJ[<#ddȌh .w4| DңN%Ik䲱$A#P|5+W/@GlvQϠ|g-,hEOi+eY.$kik` ᭐*YaL0ǃR/285t`*Q d DybM G 3soF/tNj4uGmyUWϏۮ@~%BΉB߯ZrBQw|Otͻ?lu~:zοau6[1=O+k5{|%n?-W.\Hv 5 {5>5puzs-, #%~:T6zӉ_Н('>t8=o oT>6ݮ\{ǒ/=ެcN{6lc$u7\r}}xQqV6 #W,o:tݖ[Z@XܿΖXXEkKG>;681ǻӹw{:$˖ ͣ);eW"blD.nܼmֳjح0mJC2:i(ƚ(sV5VK:cU%R l](H19ua9Rx[6(7 U8e͘Ds ҄[ 1֕2=Чfs)T.3- .x-PbSkC[{~?ow~8n׀gk}ݞL޵XC~|`'O(ʟ 2a׼xz?/3WW|/VqwJX*Sep]ƹ!g픽b΢g~M?DQE"!~]-|l8{ȋ|{ldT =ukcK4z_wZՆ/jmo&ꀿޏ^z?soRq?z aǶ2b}ex?^:ǽDtOd^,ՠa ? P%5DhG&*8%Q!{P?X%(sӶpZJ"rGW>rhVӔ3Ou+\;is[v0q,)H!*Nmr1H 5?$gA_7޹޲QqFhˉV\r.XïZΫ Wkw#iZFtI5iDp1F2c2A|_CTM>.R^,XB+CuAq$ .XK1b_wC蜪^AՆ'δ_,6=`LJOo;Q}ҬHI-O[9!ф+H:Z*v`$3 CZDf x29,*KmhK Kƣ'0\RJnS̜#Ix8 k ]:~DwzO3qE,OY 0xWڌ@ ̓Jk 2IanH vNe8e|l'.[Bk춷d E /S7ֆpW'緝vCt8V½;L+ۯ.[Lɇlo5;Q'Ux)Bd:rվoH~(Č;y\v&'Dfޑv=g{(b$GDIQ0d,$}LŨF-Wl+;:;2jy4v1 ?4G|J(!t< ItJ\Xd$4h1 0ae%I/!IX޳RgC1,7r IGccx&>\?JX.]G t{at)O!U{_CnqܱO X=A y{pNX/XZ8ERK :VJu"JW)UTRUJU)UTRUJU) (u-_*#͕RCڜthHCBZYsg_I9w#zbg;k5C\zƉs12 %ͺ:BiN 6FsiM֖J"aUyM+, ǓV8nĞ5.GIqg{Ga%AKF/?]Otl:K3Fgג_#zO?Ѐ/8v ޵/0RqllspXQ,R^k%IICG45Uտ?QRyc9o?aNI0Gx4=VUSTNJ .Y (Pg^!WR_,Ihu(mğ6DmuI!&5QV{8f@Rsу$Ιk%87ZzŤ>1&@tjLzbXߧȆicm۪Qo/q+-Ah@TMT*pCrR>gdI2K-@m;׹unzGgM`t֜sʉYIFV1pd3icP2:[1t$--㪄v?IJ=ұn|t |y3ǧF/mJAM>Np:xy xi^= x !'ƌZ阸g 3*h/dJ\F&`i ,:7f*Jݗ5To!Hz.5܎'xYBz> ;w=}{-={[6.72* 4O5;:PZ/iv;:]ٮ qnRˊZV3jݿm{eݝyvz7\>4?DƞwtLGh"7+v|;=w^yN] ]L:}Ust5ҙW&|5YQt~-YQl5?eE?LVTzλMs8P^ZM)8!VJJֵ*eR(84ivnAnB?.u=g^~43Rzout+!j9A>9il1|SV-$Z qr' .J -<ղEi]Eg֧e&TZ߻ ~VClPGMPӆײcw>Vц\l uf"8pZh#x] &+N,҉e*(i kPXeN:aPkUIxwsKQ 4RwrQLXT̓Cyrq1if݇e.9KR KQV;D[1rIF"0o:ad`<gc ( 2> J[ TIޫh w.#5 mMvU!Ȗ'E,)Z*ORFNRR@;\DZoDlYYcHŲQu|]o8:YKҠ\[# NŷƧf|wbtZy?>m^`of/K{䉒O B{?YT.S{wbC:᎔bKg3.8b'gHe: 1^xBB= àdq]ƀ̛+Bt„*Lsy@ڡꋅ ]Z-oI޽[汜}T|LJU.=p-K'lkzC9xe ʺ{k뙵M62}S|U`v2_L2|v>Wb$ Gßo&uX"p'XHӼGRzHMÈa4k^*|"jD;Xvvd4׍ }"Fm\rn9ܹ?篣Tco{EdM/e^t vNR$|>?/?~…|=8`*qMSSe Xϖ[9JCKJe˫Z_aӥ,1rI. ҬۢOؖ{?F4m M͇N3gdǸܴqzw\kԒ%"ĉ"f#ؠ¡&]59oFR[T ~p%) Af.Ahr9H+:eH'A2s4A(e$=a.]}~?"W>z7.&y#2tt  Iuk5@jfbBdocT.`B#R瑗:c\=|a<MfՋ{S9#rsBowhWu/n)7`^Oo׿-3aJ"cUALXz ]dWhe pJ%N˒Y#ˠu22t" :Q8n)1b5y˅p볱dsɌ,Q1[&8F5XBIyYya]vwH~jjTCq&Wdص!m߼~6P^f^!g + LQ "+j*4FCti4AY?@OEPgtwX:T^~ Q\mD`.)A2+̒e)̅7';`&1fn)Zn.Zٱ%.sFK۫߇R)y at ( f Ȅ/jOs/Љ$:D˔ 9"h%Spo$efqcTRalH[L{L}pR}'QFFYY%UIqW*v! ܳv&qOTa06'o\b[ˤl" :2yğ-⢀F&P2A0p],+̊({1 )$tFh^2IB\ 3#s`( F1m$C \'!&V[Xyh[9kS%V9uޏ/h86\H*Q^u4Y[K˾@ 4v%?#wbzwNABDZ]_x:c*H9 q{1dn{~rk# ? | h 39Ucf>M$-Qd Rv t` xV'&2Sz.}̊*b06ڌV9JaIGOK1!t_dnsoۻl=fpM.>suz@|i p8GfDsM ]XYXuqc c{dAô 3eD`!ॡ5 sf3NK\D=\U ),RLT{]>Y*; kySP l$SLԤ&$J250z ˾$*Ӎ]-8QŏKdJ&Fzv~>?m]^LRxҬk}ݞL"V6tGyinUGV >'7XJ.c4?ӫ Ĺͳ~{Az͟=smEVmQjr1m NP}wURq5qWwavBmnW[Q/'O˓sgz-*Ʌ3|ݺLvǓǻ.}Z~$դѓPk(:&jP VCPjGYWC *n4k.J- U*eV_|q<7UKISR߂$BP1]t%ҥ_w>ўwfoL}>Y5oݣ7aY08$ (^ˬ:E=W=@eo?э=NϧqK~{K6]V-:գF[ J~x?:>u\}?v=9+GͲg(-ڜU矾>DG)[-iU/~6]  fv#~[oj,uvVဟS;AUD=e=/m%&_Sc뫧m%e`*աUU5k+B3MX^]4%l>ʱZgofݜ, X~8X&}lNWv9!83paQ47,r!o.?~YsZ4vțbK~x[7sxF@8 £A#5S IA̍ )9b{n`eSj}LhJVӜdswz1ΐ.<3j|y0A+m"5uYU[Sx\ 1z*q8;)mIE.zp_qT"< hN6@ :-/xk"&!TG`4\/$exJۮlM( *m Ũ5( 1ܺK 7g ),a xD'!5TH"hDGN<(2*S{bÏA2}1YϢhRSQR U*Űf섅(Xh,{q۲/՞l楈;.^Ph|=:ξpV lq&ih$5$dŃ #) ph/ OC&{.2½geD((suf$7(U)@"OFbGl? /]L;vEmZw";k` VlƘ =1Jy+RzI;aS XiQ As3,(AiNYLE=-O{%۴?qnC4b皤BO2z܄#$&%S1c&-8< xƓYg) !᎘>>J@eoYoq2I.1P^1) Sy( '=Fz&ThQfPgt9<a*,#-FXKlvS㐝uBmfv~xa~[Y_*ـ'8|ٗo㻄qS[RgWAyK HτdzqFq\i"6?/ɐcAo+"Xu.⨍Rvߗ]_&`fǟh 6iƚ;Υo]w>f1_strI[Dz@O:+iȊtŶ(z])+՚ CvWf|6 #V<):Rg{#wA+$G.kƼtO=h ,Rh5X򀺫֫𚸙v#fmڪi32yr7a3ֺSv!cJg!J k#tOn)MJܿhؠZ˝)eD`!ŗॡH Iѫ#2QO3V(%D5"K>^V196 lwۨtޓ3Te31*T!0&<5 j*8BF3^i=e_Ʈ5nuOjܖ {U(ǹC. "+t湥zfq<^(w#RR3}ha}5JYS!X]I $BPxjQS x>'(۲sP˨Uq#oJՃcmB~\a4z5]j gl7P[tt t$8Q8Szߖý$x(N U+NBik>V i /nnP{l‚){i \-K5.}c1K**be-^V(C2nq+9}zQ{=vkRxeڗg/ B_ꀅN VS-,5FJUBA*%#Pt z8SqcQ9TjaʏV~lSq0B* S(aaNS2颫Ft 8F JGoBIDL` ܙ!N HЉ%QOydMKe恑dDx, &I͉<Ҝ6(Ƙ7POVph Vƽ%'%UF{D m^mR$vߠ7WCNkYRxM~a{_8+,c+JFXod9kN?yu}U!;<~Z}d%շ?g (:sL|t򉜓W l7ާJzeC7GPEan|,z`.Ҍ|\Ga~E2IŝFٞc oRt[)'Gg]}1IJ; c;Wjnҋ_?yU֊vX-&INlӶ*+˝DH9we+ )9m ]Vڕee7*heK/9PLOֹ A{Ki{am/[&v>y] }%GAf%{]:Pd餩u"P B0^se)F˝6I$r H|JQsP)c$pP8qxpV[8QȄQ\P`HbYlGdgJn|rɀfW˻XqMoC>,BMgTKdD"B-2QʻLPb \HI;˃9$O$,=(L)L4SUM"c/ wD%"n< ^+c7$x69fݨq$iE6(mbL4Jz^_ω{ٲA?[nߗjd#9-N LO!$# `'Aڼr%V@ !ٍTؾ2f" BR+\v@#{ ㅤ UAӜ G0[Åy&IJYZ>KcӪ6ld~]?agj'͡l@*7KV&\+Eŵ\XBH_?NF7AœdMQlP% ;듳JrN?Asg! 1_-IsqzBJrrEW8o7j(*sQI*/=bۿ 9IFlo''6T|sȅ`w K34帩u|@=q#_`TG0l_\=ax=r='W̔;馛:7Gl^.nðJp7NBzR'{v#H=nJ> YZ|L`p2 z+h=uod?{M TdqNhyϳF+7EW.]Wׇ~x2}u˫?o8s: ݙoN5d61Oq5Qnfsϗ~ΠRd,嶏ٱ%u&_AF&AM,AaZMOn"pS]c]S" 6&\?$̘߅JhG6 rT'jN2դY 8?_<)n5N Dy) [ x*r§$Fq p𔯻 !i߷ RV8I5=mxQ%\. yɴz)([՘ BrYh#kT5 p/h ׉f$:rD?@G4#@9uIcTyTɣT  B2RV g7г"zz$w*˕` }}dC>G!qKAOF\er9q5GՎT9X+N]\ >zZeJyXq2j5;zv/WW.=ŅW}{?l+d뷳*oWwk6pep2 RR\Η.JK)E;oD5u R:g|NM.y*%;VsB NFer8EQҘV}p&guT] L˻/C2IPPj7Ŝ7S))!MĪ.[u%D=&&wgU!|t#]gp?PdIϺEt\5MvArټ}h[M͏vhgr'j.πǩ% TTT?a.#̓L4GjqLGdTA/ Ęā3JLzy7S%fпYςոato焮e/LϛV~7Ia֬(תr)y5.oa_\6@yZ tQRR-fV{IBw$V5<#@>bNR2D *j2N'儡d൰Eɕ 32ֹ#h.P፧eqq,<҂H #g} X.i}ϓusW =BYW~ַϨ/_0m&jlrfvaĴц _Wy.~P ƗB|P,WVDTu @XmK'A┵($wLr-' (hR{uAilerj 7I> \ІH;c$:'-u5 #g4J5D+R$mɞ$̗YNݤ$4*ٽ'`w@ *P՗EBi$*qebI(Pz'$Vd^\!,ءky|G^Zp!o?jYupLɨ;.vT<^n1z|Db(\9뎺on[S% ӾVNyp Lܪ|Aө:c٭ՔF Kͩ8R@JR11MecTxV2W-RBU*jL鄔%hJ Cv7[8S&z̖SsAqYNWnIKqYI 8q.FKP%818ccْ3JY*ix@ ŵ@$?{Fl G`?d3; \,&{*BdɑL_-ɯeEnrj.*OHhtA\UEIz6 cgs\o1<ܜg)=Rǟ״V_7[湄3i8qA>PJhSsDMF򦯔I&H*;]Mf2$x'Z\ZR췘UXyyoϧpsq>}"uU]w[tWgdb[!Ojf%^T >46nB!ج6 ~_Z2ڤCAB ֕Rr:(BVQ-)Jd\L hR)ː5NE*(h&gfLJ>S.'u1P'A(z'x_3`UEˠD^R bR@/^; z^FtJRqL2-DYTa1p9EAJ/Q$(/|2ansLCH]PF&A`el{ugÏ6t&^[^4yL{v~wax| /ptn;_,鴩|:+ ^owTޅwzvlY;kz_oY{לJW]k}G Kx\kfUz{<^ٮLilrɆK6kiݽmU͗|d.lK}v^yMBmy+bޚNgӛƨ[wf+eyݮѡW5A7*\;/Q?لU⹄^W7tR_YI} &PFb bS=\XF >K6,>M6տ5)s_71=G-f[V}?wIPxA]A k{*DQ')J4j6em2H$!E}ڐER&8 гEh2kOƹ᭵=w7:tݬ&7BQ]&P{kVcțExoˆV҄(\JLƧ# ZzF!5zpǚcJ_uEn"0\ƎrsF +f"w";)jSW{At[c+P5W ݲݫdh^ )"$~IE$`uN;$ZFuLV xADk Akke!%Xe$Pe6uU=52]%!bR +hE8О`q2;R={ 2)ΐdT5j^ a-Fy٭UD*/Fׅƻc,r_'y|WW0)qlxQ-Fǵ$N{M _>-eh ,AZ eՐGҳ2PmBI!PGg蝳ZKbAir~X2$4gAWdBP$MWt{wRx6oQV)r߾""{:Hk淫Q^ |a|0Bk9eM_ ˈ#j2ۋaG'NxKQތg]z*^ODT/탗'^]%f_w<ΜiߝUyb$Mg_?.$X.&vBӿ]-Bi[:jFu6cD k^HYON7ls|:^nU%dWc/[]3Wڬڰay6oSE|\hVF2VOU-Z]U6O Eoz_~_|/zV*__}Kq3\tua;^]m|=A\N:qSѯ?Ƌ\ok0Ff=9Ucg# tub0^ۧi1-_M_14iZ{i)l^J;f¥Sl!ī7 fq#ܠ=*L±&f鈏u+1~U?O-$1"5j\6{QZQ2`F(J:at/RFAIzlk]v|9 DSx7.ܖå&Ro4!Dŀ&/oAB9䐙i3 #JddV@- NB B@A^%RlǍƘ8}'1&zp u7c_Y_#@EE@D`K!xw ^0kNEJrM@@YLB x;Mkxx;BX쵘Ay"ç^-,8ya&Ss҃VD6{@I1Fl  ՄB>VsGj&{ZEVN/-J$W%Y!^/!n!Ex1!mx`iK&tڔ\MT(RAjM`2J%.nP:|?P:ikFG-RP M0}Ƣ ?Br*[5.druu!!D7d^_Ng|BN\)s,֎E{$y %N$p CB%.DZ^%d2kRTYfċ)mXi[w^n3-ʹ(pO79d/˷yV~#7QRXx(/*h} BuF #ɫC8"`{EF]A,)[`g%CJFB_ D^VZ4 T[wFJoX؛d셅LDX,|P,ll2k;GZ/rU&N&?&;Gl &hPFS (QId: eh`%/^i^D%tJJ ̎d = Z0OR} d*T.}unN)^Q1{}Q[+0,H"" yz+eLL bUkcsJu-ƺ !33df$)kr5H"[EgđjJ"RM.ƞ܍RXP7UψDq[MRhiBY"XgE)-;XHBX’-CNjnV3 Lt|άI+p:: ^%{O_ub\\l7)mϸ\pqk<IG#4EsAB_$ Җǔ|64R ~I~ŇŃؗt싇g< " u-K:mȝs1w{Dޏ{~}6WroB I>IHRF&LBS hZC:=ٿk=nlegYANR+tV\rJT5G 6e]P%!G0D̎Se*iϠTt=hPH,wPQGa1zzΎNss6}>o7C <[f}Li{O2NN<+Oyl2?|v:Z-vlwebٻJԷ}BwݍP^ ϧ'H[݋556ƥo_Nisݍtw!yq$5{gWZ&y~ۼ1|'ly3lb f:k:h&;ϗ,.LֽI~[rugQyX|Rv{N*!r$5VKq2QGmW A޴މtOg3PЎ!:zgӬs( w6̣`3&QŬ{ƄgB]MT.-BM/:z}XNr_ktۦV{Z[\saPrN)ϔE~u'.ū+cgmk1բ/~FYW^ʆwv*Tdb4~p~)įqJ>ZMwK,+㨬c-qLUiofXr]%פ Y.8q eNFD X Bn8H4&ܫ[%Ɂs[&G<AesM[3~k9[?Qgo|p?w(jRHi(udTdWi^hRwiH^ߒrT>y}܃.(A.[XglRhUeLJ^MϚTY_ӅHI-O2:!ф+H:$LB4nv˳4ˏ{C'Bf]ɫ%>.|RbOjQ w>=Ikෝ0.7~GЍ{¨%=@2CC0o~PJI/8MoB1D"6tJH]Ԝxg[!'Y{0C'4ZzL \F+yTK ژ`A(wFOKo̢֢PKGe^"&OFuu.;#g#Ç5rj&S^jB+֔ Ѹ7{:_.Rgb+*c…;CG~#SYCC\|%‹r:t |2JI(AKQzc'`$=$]ݲE뚗bp|͍xčrWwZAjLwF[xo/gXJX FYYS|/_PXeM;*o$Wr{BӶ:4}a˴U="1/OX286g%ZP`ctN=J{H\JqH2JR%&S>/SPJc&{7J;)<|:Jt^Q5ϱ)pB`ﲯ}MْY=$ϥg8#CQҜ@( $hrɺR)PY$UU*⭜Ϋ|ZD8ΡXC8O3eb26()q3,$'GK&y kp5A@_6n6Y'98wsRX46pa2' N)T C wu>MBL">7l qD *8E͙%9#; 7>LQ^ͳ_H9+ypzBrw6nrczNny;5EqX]T}#b(,ЗRw)^$ղkZvQAEv fm$G|X ω3<>tkC2~=񟕩ɻYe`v\FXsa8dzz^9,uHT GUh[d 'Z{.ںaH>eFy6ᨒO8y)>g=Yy~q>]38Ieku6ȶ^[*Y:n8c禯&"czDQm[4kU֍Dž?zLsyq5rY.ॶZ]#PeUFZe;z;x^꽀6hƘB:OF +q)w)wG_!Gc#>߇ՒnLr׫߯zfHQ+rw đ9bOUᅳ8pEpJ&fZNȡ@Y"u2?Jo"5_}0nM}y;2ע}3;ۦ׷c>nfKjZozno2%ZR-Qq;-+Q{%M*ʺIR LV+Ho}O5v|td{ *ߦ5lGRH62l?A+R3RKg̭ևY;siY mfߗ.6a>%iGJ f) ]y.4z@5S;AMAr1 !W)J+u4!&Q͂2H,"v)Bx[.dRxg?TxN\BA5$8O@9Q^Wx:D9r'EіT$@&oI_ 4$qɍTyĄŠZu"7z0H8d`UWYēxzdQU9(j{sĥqvęĂs$K`8"IH(H-(jёO*j)=əDP遐ᓎzDp*(%4R-c1vJ1[XL2BJ--ܫ-|qGѿaȤ",oTn:>\,?s& IOI~r"Ԑ<RD˥NY Ee,7< \q&W"jf#rh9AbU iybUp1jVËkVKAeBD-ԋz %~+Rz{M)xcu,P K*CFdȈBD A1`&([-ctA;Wcq{X;20E,&C-"+lhGSim+% ,$>k;բA%!:ZDk(!J#P 6)M9NR3>z(9ГfBs]aX/_GXYg1)jUaG8ŝ=ۄQM5Dhhm(F\7{d$Ҙ?Rha_)jua{8g&`Q\Vo` LUg6\]rwtŠA^z$fʑd+ༀ1K-`zFJl:wܾ E 3Dr,3 es4ĵeZcDpBCt˞/3״}Y]"3܉aQu;~?8X oN !wĢ;-gҳaV""dYnx4 ax uR@X""Ӗr든 >19S΂Tc"Mt|xwκ+\WaH>\(6;:dہ#/3QNiAe!6yIg+D)0`ģ3IB >(`kI@z4<2<"s<;b"b|4(B0A[@(}fEbՈIaM/=v#Uݱ2L"'i>{7Y\/Ns|-HzPֆga؝]^g3BCgor &՝0r*dlrⴉm3\eӻ6Q샛vͣC0b6K&頭kYӃ}%9NdLUfrqlA<3%gțuݾ 5RR!ثc]R,ֶr\yz)n"nx?|-ubMK>d rbO~m9 Yd9bqǤ+9FѤ %":F44[z2)0Bi⎝~I!yc=bmlX\.>A٢ @b,g47;p/=ҕ\/?vYj{Wg ԫ'a\51)oK1E#&m'|f:u7=Hvхy!FYHyƅ T&'r6?lKxP $դѓP+ɨd~tiк~Cifo-V讳b//`}pOs Ep<9ʣ bnspFlҹ#9hX[#IX%M[T&o ,ZGpDYtO& f.)PR ;!cq ~g}z<͖l=c. Oԉ+Fϲg)zzh(=6M^7 f˿@L&,=Ȯ keI3I[>Rԇw>yÌgo"4_F V/jt,ϯ}LIcߣ _4Ih&8vu(qr3{?_Ev^[t֯U@on&u&5(ƃLU#ÅZX\3پ8|o!'4bdlr~"ui>icWM|M60/G2fpNտ,'5&bj2a9CNNI +䷵8v ^}TK?9ZھգyɏK@Z9 o>X+ ZM@/?H\ 7ZWnD,>|X =;\҆յ-U+iiԬ!O)[Rn5?"Fev8СZR}PZmĭ6ାl'nbkOu<ߋ87clivv=9܀<`ZrZ"Vkn#cqd1h5rBzW,ng1Lo/{ܽ!lψ,rD}7wͮuSʈ, %I{'_ȿOE }G`dHACY s+i5 FGCAp޽GVȶpɡL.22(ȃ7`C  9'<0XfgI)xkb& dEuLy+Cj?=zCcOg[6>kJw'c4>ƏA_fűGGOCGq4$kDvx,(̧ F]dDj&#>Y{tŤd* YJx9qkJ=`0@`h uoűsod_o|(ݧhR ~|X =z B~'GQ~';Aby ̅ւҒ̼u)rj SZ;@苉>E߮ݸw6w3UH_ۋխy.0o#H[H6&1- c#rZsoDZf (Y"׍i_Gun%^U,Q)[#pI;60evF;"#:bJYwrt>n^~K>揄edzEDoI2IjߋqQ 7l͉5*,QDAֿN>Hz!*/m)Ft94"H+"UGw%O`ȅ$8!@L ,pOۘ 46֝-YfBhv̚%>#X_Gnb>462LSP *lCpLa˹}'Ӕ4b:i0J6lr:S ZN+5~n+vxW--A~%aR_3}D $Q 'dpYq=IR4IZ<0!2(ph KR\G<8Kd[^ \J 0xmJ5CEfu2G[,zEvjfa_B|Hs;6~8vmZϻp-Vocxu\+9yJ?]kvڹB1eoeBsZy5onmKV)52@ܶp;Ad8ᆜ9g\2qrNδpftGOfs2'idF@E %#i6 j?Ӽ eExFpz>q1ŁԝK5$mm3#ѫWI/5dI Iأ 5a!;1HÚWNoaT`k<҃EH[u\<}|M`ug̓]w}hxvWiuch0Y ^'v?c]-i-]t5#:A=MAyC x'2eGW=nsx1rlUݭjXVg#4$>LCj<(&㔟dq񿾪ԢQ"cQpx<sǿ)z~Ƿ_\7H+_h)d]]Xݫ ܿ_Oeۺ9ð6J෫O?Oقvxٞ k[Hz!YA\}e[P giAՄoVުii9S5G=;vk5!FYbfk;ҍ L O^ ,Orz͇Be2qL4^ (΄_ITDLX(c''|%1* ,GDog,WȒ> ~MN^H:VY][3]xk23 >bƐh|fjZ]Nad;-0M$(|!li;r3ny̌ߵ1i n1O# 3Y""=ƍh^Ҵ+x5Q`U# 8hX ̃Gjzf{5t'kɼ'3IJA"GxHLKW֡ di)=? !uiIv|?%'b}A2; r0jS*^|,6 @ۜwh07Cn B_m @xL^d`=_6ȪCz.)hM*$EѡV0ep'mHsYFqZe=/=y!;%0DvsuIIT֧ſ'e׬yX)Ӷ C'zJ'`&pͧhE Jra3k)YPbnQ1emAZc {/6Yx?Mb#'BZZq Lbw6lϚb?`ڣ-*7ND$a QdX4="?mSNEBOѢ&,3=IL0(BҎ|uƠik(m@/z>sS)s: RfiRq`T1' J31'U5VE\df#g姘<$$ZH$w̄ʊZ;w+j:GK&-1\&'^zզ˷_pf|Z: J˒h2N:Yr *PO/Gq!CuminKi-Ġ2=^^oˠU[}9G뽶Rب,h@Xcd8ʇe ji>hKq#,~oI{#ykM"!%aj%A YRQiɗˉM׆틤+1>F fIrcd-ֿ=0n6q+.?[qQAed +c69>8]^9xQgvkz?T{SA.=91)٬"jX^hى)D fx,Zc 8@8!hQ) !eɍe T*H][s7+S~ٌU~*:IN*vj_RjD IɖO?!EJ^it`@wcpHn9k:& q9l_M鐂eo4Y] 8)KAҎ_l ٗ1EQ jE" |oT:@ J%Q8E3E B劍[`hlW\F"5F}`HDAH.AMH*E8;I]^)W)TW0aX^M"xW ԟWwzm}aMjpkzP5rBFRk,"5y}tAn"O=hBf|&~tA~!NCJovFD-kԮբ Rӛlw{TNXRW#C[Hk@o0Pu{Ecwl 0!;}*TŮz]h:48kptMz6C1Ş<YşP=@ڶu,:]@~..U*dZQ%UEz+q@NbS\ υY(/U@; Y.AFj)mj~LtqM[_wwft`pOc< ^X?ae*}{o "=sbJJ)v(sz:aaBvUp$U &^5TZ+xAjL٨dUHu-ihW Ul;͚YfDDru( g+5{ 3wNi:P%g(z42, u g4bHҍ5򍡆8s$Pa+ؔX#uZ"ΠA%tS> Col5BHֈfƒǪ#@QJ~ۓ>݇a$Wn)v^'E+_i'Ru=p:Bo`\/޼3 Z$X>,oj۩ 1Ə \%r59j9:xJTla=ծpV#]%r>j1>xJT +ʸǤ D.WZNW )5kh!y;)S^ৼv ^/gmQ36.+o1x՚ ZG"3f4j|5#,RFXp4nX f>9´L]뇋pbS! Ӏ>_~6r3E,>xa쑺/e_&zW7&=Do{$M}R7 6S YSJ}Usu\7t@ZcڸbB=|_9NyS:'iGsʅr Ვo$E8' 7T}p3[^BlIvyH煥:2CsF5/)r콓4}Qb)Pʍ"VFuYqGkZ D hl:bc䬎!!f$L*ɺ__qBKъ$bRgx oXʇ ҄x)d$H;FPD2h$c& LX+^I.^/ K%!iCHpkļV uSXz,3)"ָQR`B<UX-ƽ3(d,`2b=6 jj4`Ѱx9k$v`mxxW[,U)u.wY٣ɬdB{F+I :Uu&!fWT]W=unvHW&)z@Rd$Ahu*nȀo<$Q*Ld^n|BiVlx|_dngJ.`ՉgA?fYU*}=b֜tRtyYg X7W]V?D3qfd%C; dKG#L9F'SwbFpe{w'1Iz6 tuY[lBVTa1\uU6E?Nc,>NgFsqpzkSIQד $tNWקQNǽщ[ 47Q8:QDQn`# Ũ/l?U,NDT{]lV0{L *|n2#r7&Zw.ü!m1p HFu;<26]/JzFT) P~ V  ̭ +8'\$߶ͳ7Bܧ!zVɡj6πQ^Ft}'!@qx?|e|8>eQ,9b!;ұyn+(ߛZ7۞viьrӃUeKn"'Jќdsk,QV8dn9AI@[ҳ9 ZSqbOōk]@D0VZ 6r`#SfX0Ҏ#5bR,0spe~ oLNT&ԇ!Ԟ]^0I &uD9$Orm|̓Zmw6do}u BƤ!fH ^ wE%K]mo#9r+| wV$kE$"fq| pX86Fc${fg[/c{,Y)5/3Ů&"T>B ٢v@)}:jǍ׋ȴ^i (\ô >؀e"`.RtaTʲ#Rk2Ø]UfW޸1& KCɮ|,Jl(Y" ..|)SA3d9ȅܘ]y;diS@7{:i>^ $Q.F=Ug|8e8|:Lh{DeJaͼT|Q) eA,\4 ɳCEh"\b6T =*}r`T9PJ`BajyqfXL3B&J1∅G[7+3RgE.w.LO7&Q(pTDV'0\6,XrYu1 Y Ҫ&dM&;%D-')ò-`>8#v]&PPvڲ%j2kaI$"),FMuRJvj&RRrg 0H23CfF&XI$Hd8RM$BHlT 16a<ơ b38UcDT#"xt6=J% ,|"T۬Ylu$[< PHh1FK,I͏UL̑Lt|JlI+tp$Sl#Rx#b]dnH{ʹP\qQ8Ό0IGL)xPiK}z)"\d~c`ҎC5~a +17UrBޏ>vFM ~?Fi!N΀.d&1hh -yNwv`?_rd=NUvVTrjXBPDM l &0;mQQgk(g*Zya([B* ^Y$|E`c8{:ycH|~QR:czwYFW۝dy-TWyZlRrySG9'E' )3$K"z,>؈%F bWlVUWj< / &PEPjLS,*dl@:u+Tf7k"8"Wkj flzwnfoD|[=y,dXvԽxd5&ccR1YRJ;/Fi, XB6K-@NM_7,o}#;Ik4D,"82Q BD*iXI$RH񃂺Z|{Ixlz>Ocuy=Uaz sm;(\_j׿s>ܧ;{u%&ܷӮ~bۻ`KzY;m-q`zk#:k%&s*SJWeI )Y'f˚Kg6H?'e>0yU+Mݻer/ޖP6+FK~zsCKspnዊx][ksl:S춺!6 j_MqP%\~%~3\7٨$8$ڳqM]bHY))Փꈭ]Pk_ @~6?ǭ>N8X,e@[#Wugv${q9t ϫw9Ndx+V荻noX<{k\# [G(l]!]Xb&~wXMiAkS~GNm~Ƹss͂|:|ow/<_g+2qj~^e|pL8Pd䐬ֲmMw`l/1txқ!{ 5:6kp gxuK3[ؼ/Ҫ`NI)ѥ(脭eRd ^Tu Jv384H޶pXu!w :~)BƂcՉY:H"0§yXiSljm:dy:LjfǪw|_BM.w{*My`7//Whэ>K _%w$òE]gmZqF66HW(P&V炪dliBtP"e0:1[Hr

4Y06qY[-7o]薺ԝ5Q:`:LnX\ +߭{w^MapS+\+p $MVcԖ=}b`)d)R!z%l9lf-Lk)2ui[KWؘ4x@Tr6;h~v21@6bc١,En8{nvlt@!~vc샬/?x#nNLuWMΆ5b$gmyrPgG1?( h\2 1"٨SBR.y F_$\$cM7n" k+,I4dkc*7$jxz͎RI‰T0deG i 0h4 KXK0QJ:k0D _|f Նh*3x;ҿ&nwlw}\NǗ,JAt SKcV;6Cc{fȶ|\ ۋI2xϼLB&%4EYo&+S+R.h@Y1fhc~ɺ f݊ h2o[k`VM+\oOnm]RkWoI LT_7uG$;r"Ykfk1&j~9mA0Į(R"wt{ 0$sLN%Z[ճ2OZEN!=c.}aһ_gD#M y.&W|_]?1g#Ln;9mCx1t%~_;zdOfPӴqͮFnQ-.ucEv.#z!Jڃ;U`I?Ч~J;UJcW~ҁxJpKP'WU\O(W,nWWXyBp8Jq*pU58tRZ5++JS+آ=bqKB 7tS+?cQ~izƼ,\=NZxIiweWvC^ZF+BpU[y2p: W5k5C+D=++pUvx2pUEu*pҢC*2#\B8%vU~\ TJ;|vUD1k+a O"o?ďU%/wQK^I`?MfбI^<ٔ,ꖺyƔdw}W/&o?{Hr] !jz 0xAl K&0깫X/މsIhzFwD6٧=j sF [,\ W^]սbipF |w[/A6 A~DrGHz49%A&6ٷ$EŽMyRZχXÔcQsnSsٯ/Z R e *qww|I9SbsZ6*vfnBss$4Oee;Ofn/bم@x쉥E*7?}5]Z1\gg\3O!O?ݑNhjYM:~*S iߧ;L-ĩ@zBt%C ].9 ]-ӕ@W/\p. {K+TJƽ_AżDQ ;] 0@i3 ʍz9t"0U|h:t%py2sW`_"]< œ+kTJ:t%(9կCR 8> WLW:DW=@W@Wz80M8~}Ƣ+ .F+AKnJPZsHWOsҵZNWҁ^ ]LqBt%t pRS+A>ePzt"ǫLGj B ߾VuSivf:&XONӂ_ Mɻ`Cv2t%py2&Xк+A&%ҕ_Ni=dBWQA] SJ0\;}+Ay{tcR+d p'tAZw.WCWaˡNJowLvY]! - ;U8c^kVz(ury~~y!A`t@\KDqe?/_zsn6zh.Nz8:^bi=z}7ݷaAλ)K80lqb$HoZŞYُv`Zv2쓆>'V xէ4Kx WBHnN !FE}Gh)Td@߷G8Giw[ z{^ZqAE~̱rQtg+6}Cۿ7dẕgnZLsT+Vvtƌ4n_of_32Q I5l#H5A\V%)eRJ5Ci7!'CWNВwt ɴ\gTJ>תPzt*2Y3S֮ VSm?]ژévdap9*SviUkt݁YF S9m:(A YlcS؂6Z"-6{~1Zʹ6xmif24-pi2י}iAohд=i'_ㆶJdV0 ZRP IW;?)rDW.8]ӕ:%G{6UvD3[jӳR՚U-?97oVA!,'wgAk4:(şϒWI:C :d1Ћ7Hgg)(Wwh:Ƿиt6F\,wlEև$[t oqtہ-Sk?,}t ')%tv[%U~Bܬ{l,P|7R/ya|?ςon{GGy|X-yBx\3wͯR>0>b?h&G^@=zdz@nڼ__|-/OXwcKC%zP5Y]M.Gvrә.vvE'b3/ݜ_$v}]t՟ҭ_1_?wy=?wp99ӓ.JE]B٫%S} e5F6BzC>2H՘FTT Xc}ʹRtUsT{C!4j~\hhϝ\}" TĘ(=3dRWY-Z3cvc&UZ0Q37mЌ}ԫN=H)|6n-v6r$P,nmY!Vi}jys*ګ;;SSI{ Y5thN1HGQo^G[cKԶwSt-jI:5Z%GYG7( }m 2FXKk|cU.JHQd3{ÈX8pDڌ^QU|NӨuP)XU.C6+הOPpNƢ27%cfC) XgݠTGhOMÂw_KͺnG0LcmTBiil")>(n+)jWH JjV\M `6+ sGcYBpPk@oJ')CiV2TSR f XM2иZPB]F@+6CAwVrE 2nPư)0u.=1ePDdBEAvcŨ3| ߚ %@ aVv}uT&];Ĭ AཹV)(J8E5n,\ B=ʃ# 2|i63 bM=%DA>h۠C^s` Q/wK|̡c8#LmDT2{"<ѣV !nK)2vX>赈 )-f!,-D;f broQGj3[x 3Ȥ td!d0XT$ fjbJL pyP*ZxwGΚ ת0aTB]%iYfJNXZb~B=`̐ڳ&&)d4ά$޺[XVʥx/GO},edԿQn$m}+FT Z)mm`=Ɓzy__t9_lL{szy֜ę$+7,VA[.n[CiA'EEU4'owG(J(u6tk*F%'ՓFCoo1PNzvx4l}QeFĪ!)a!/Q[tlM1WTs]`7"2Uw_[hO&%\I"ݩRA  C@6%fTi 3 C)wkƭzb+P!>]Nw`@-!mU1_汑w=a5_"2ԓ{XWj8s2\vHޡx/yu_lD`tyY[mGdױ8>-ge[ wM=^6/hfX•Η ;d<&gxͿak?r+-'|\4'ΐpm3M\Xuԩe9yOhR+NZ.S2\a@: \/1\9DMGʐ W[pE+2\ Wd"pE+2\ Wd"pE+2\ Wd"pE+2\ Wd"pE+2\ Wd"pE+2\ Wd"pe!)pUa+DiS4\pV"pE+2\ Wd"pE+2\ Wd"pE+2\ Wd"pE+2\ Wd"pE+2\ Wd"pE+2\ Wk.%t WuF]2+zo֑ WJo;pE+2\ Wd"pE+2\ Wd"pE+2\ Wd"pE+2\ Wd"pE+2\ Wd"pE+2\ Wd"2\)CIp0&Õrd WY WG+\ʀwd"pE+2\ Wd"pE+2\ Wd"pE+2\ Wd"pE+2\ Wd"pE+2\ Wd"pE+2\ WG5\w0m^bF/v1Ħa[b,k{ `:kc2%tasÒ񇁌0;U a:!e2tp5ύp +o&!a+Y[lՕSNtЕd!s\U+\avh醮ڡԦ_t[Е'zjs &2 ]!\R+DX Q6[Nü qǢ+c+D ҕ\]`ku2tpjqwB)ҕHH@©[xĵiܠ͎?5Lu NjMȄ0=r7xr}h Os9GggPGQChd짗km}B4 4UɨJDkdiQ6zjӡi,!Zd  ]!ڮ8ڡ*O,cΧDW9`wB:/^]!JO1)KW%3fh;]!JkNfpq}*R/*+L'ʖIuPDC?o.f-vSnؙӅRUPgʏAI۞` Ю.ru9tLx}=oQ4#?.q(Di@vc~,>//t~W{3aᆈS:w+ueR+5* B){PZǨfǼz"j!".LiɳUhBUo̱A,/꾮oWɾ*E_fjDǵd\*e]Elf*SЅJǹ(ltOW JxLB!6[ɘI%F9P[CleT\I;RA<@mQn-q}hkM#\RiDtiQzE4}4mJK._t(%:ARZ ,cl*th;]āDWCWboXt•h:(':A J'CW7+Dy P*扮⏬z>d윮]['BYG_DۡZEtʼnZ\3xBtu׋ u*th5;]!JÉN/< + ]!\#S+Q}+DٳЉ^ B]lJfM|~vpJmWCi%)ҕUџ)d#RHano6˭2%F& dT%i@ >A6N}EEWWT ѪEF] ]YIe&B<uhe1S+gQ)`ky2tp}2_DO@8Nb:/'Ej ]!ڮP FCWU/BG}Y;kkw;ڡuM!L] V=tNDWd zfS+D+L QJS+c5AYY7_2~ڶ.Ox= wH}};oׯ<>%~@_zt5uko'?Wq;>柀}Y]by1zϯbJa>f3W_(b[ ]X(x% cmεrdbs| 2|2^ev,cq mFBBP%ҕ(c^zFJ]L`.+ic8.^_k2QCͧݙm5(67:z۶!i}j:sas.^MVvul9Ymgk)?_w;tr|1܍?@<*9[bqvS_/<S}dpLM=7&`>&u%"2j2p"J12;K̭ elXKJ99+I sHmT E+qhpTdAmT]^fΞ2Д Ǎ hA^$bŐ7`V룚19k_ pUnz޷nl?ޘ*NJ76'ho0 P#Xh& ױ8>ba$A낐ty3A6׾ *KoBX ;dU 7pg叜-B5Mz\4LFOQs*tA'A6 t1.t9釂s<'^1 ۑ#{5 <\aXo y DŽE"+-24IΜp9BT7Һ!*S QĞ%D}"㇨,|EigFC,Y!-EsaOWW%wo yQr|.z !uTeUn,FVGVGmuި]lf[i(ϝ An9 _.|1,j'!{{]S ;"jGzUKk=^V l v >dZXVGP:*Ue>w.SiP. 4.ZTSΓ+Kv|&nqȢ,4(0E% \k ZyQZʼd@(S>TGj\:y`(g>2h0W2P9Sz]p[F8T;GH/;lwWYX*7+84}i)8//I:^з:GRG#kt#U +,CD: ju(v jZ]0cݖⴼIo'{x,]U4k5Oe4 z[4&XVxq./kyl,n] \` ax _bvd,ݳ!<~9ܱunq:6!DMX^>, =Qx{:7 'ر=4ކ:jRYy4j"f8f@S){eKUFxkU)ϸrgeU:Pgmq߁?Oz^½>1pLܟylM^GRa"`Δ$5dD]-d;=A)H<#rX>?1vK|&bQf*YI-J!v)kYFPBe鍭*TYBVBY9/`+ .c+Y㦂HRA<ҺOоR@UF) 9*QX&c{Ҏ۩ NrA0+i ܗγԋ?(RnYw{3d^&lW6ÈcK[%`2e&>?{Fvgylfgs=X`K1"Kjɲd $qT."뢅pIQ&zxy xO^> xf- (餐>PKDF '-!PW g$QKK L <C!DDy!Y`U^cgǼ%s|JR ,dzmñVar}^fL@Ixk9=߁Ԩ#vU43n-]xwғYCӯ\=Lndx]u2L:S`3d1^o9u3ܺLs\o!lZ lb [VNv>|x;<|+-7t2@otݘw8m{˳%0nMWJƔ_Y"/n&-Dcq9ss=4}^>WV;d}rm#̊li ZhW}%UjVX6*aݖ!XfNێ݋D ?{}Ryq_JPk&YL2yy Jڃc LM2LJϡiM\&hRHB["wϹIQ-Mђ( Ѵ^j[!癓apKis Ch\< ~b͏nlrMgIr e3*#;F|2i|2(*|4H& ʊd@[n4E]!b`ZZD.Մ*2"X)ո"˼PNOFrgptZ{"zG%RmfAXZ)'V=^ PEq]'ꕤ&1ҤDf]:\^W9*pW ֊YbRم1N"Zpb4h|"F \Pkmz/`QK%Yܽ{8Lu8OwhwUjg E嵾v~x5;,Zr(B ?\3Ey_\qķYNn5;1?R۸(Ex7x#}+ΨAc F 儝]ZqNy8su W~t{VP ^VſFpg8 w@W6]9v-0\f.)Ez_'.B8O \F4-k^ˏ~T~s`w ڏ%əaˣ6Ú7O\Ӌ;Ք xփy|m~?y*0{ ya\q _ۻy]9'nfM8kĎ` ':{r}OW]ݰn$Ъa?O!a0I2LQ+ eĆA^_8 UśMOǿ~ 뿾xO g`!+v`W k+q5znw}" nH̉h<bP)ԸX R9C:ddֈbH_m7gMsQvUPkmoɾo*[kx=SWn6e&jDbJ8+STL 0>!ui%z 53*P+|IHo=9QM#TBefhYL6nR+u1Uپt)&D*q_z5" /صPD`h4}aV7iDZbGp4m^I +\i}J OVSvJ O㒓Ӫ PA~TZFŹpQm>?uFKlJ f) }v22{:YA ;T1 (фxG4 #9\ڥ Eo *NRre?k:rF!&yͩ*& -1v^t ڲNdd8.i-cf\rcU^F51B- Z0}"b=ɞpd._gQCg$.%ވ[#$S&YNBjFDjAqu9ԪQbLޓ\J]2T=2|1YϢhR;Jp TiX횱V) I^) m  ^dd <cQ=&Fzklm0 M Nz$M''"I 3@ H\ʑaP4Z[rÓiƞpgYm": Jx%f6"IL /V *VOq<]L:ڴ $O,K"bKqp#n\J/s)oeK\%C’ "28h5HP #kd$jևٮ~hbFd5"5b7הZyp*3$Df8/.NKu_bܕj8l܁-9( 6iG9ZgډXΜe"&hk˒ƈT ؗ=_fyI@x4uS2Ap&q.[j(ΎNSBu=|m6SJ[b}~sNy2E? s}s܉P/Sn>6})C_>*J )ފ|L,ʽSrP[p&=K*f%5A4BV{+#(ᨓZH[ lșr^',FSkI9`t8Ӝ iF$.w+SktL>'YUa yNFf0eb6B|4_ Li:|0ͮF9wd~r>\>O.6E1)$MNcS7y yy'+URL٭}EH%fBږP%AJ\3 ].6oS۵P|!3\7!‡jҀzuEWwq`i_'"  oӼn'a.93w cu3gnqB+ꢻ]i:6os.z:jv,a` j[m-!Kb&x[s11lZ톎|aCɲSeL8On=Y-9}y+_mY`!ॡ7d.@p0eQ m(\S >%ޜPkW߁9lQ|ch• ،IT11੉HVLG@`zWDe }UTƾxd- O^WR~|GJ)RZ,"H":+fiXi@X |H+l-*PqJ@rc+m_QrgHbV9ڻ4xy9^ՃE7g[.rjG3|%yo%k.˞ߕh\(Guh,9)yp+DpᄣvDo cHP))[Czch9B< 1EmCI!It bN 2)ݽCzv7Rʕ%ba8s 93pȔL'&ſG4HBY wΕA[ɉ$`j%l2=))MD  1/e0j0yFnm&d_Rrd+-#-Qк[7 MD3 m̡Ce.Pg 3uۭ/mG :Ńb &ԆR_KH&jkJ벫e]ق<K:1cHA뜏 ?@ s4 f*x2T%>EU.yF(j.d"ޢsۧ{ߵ =Bv󋶶> E'sU[풁qʣ bnSpFl1Z*Hf :-ie⃥2yU`)U:8b+G6y> QkS ;[7Y߁O'yڶs:[:vLzm)}g2%%R^mՅ}abfR3 LME16´|,8)!q_\qΧwl4Ass=gA㟟2:y(dvsԋ"5j<\=kpV.k7g ֯(e!/EFmhJ(!D|z5Jo\Fs0Do0(|m^HAH[OcDD{䃬D U܈缭f.b,^NHolԒĀՂX[-EaNegI/;*S39b* EzYɩLh TZbk/EYVk ,1{.YN>H5Za:O<0pV3˽&Rs}-Uu1չޘK10! Ҳ>p>`5}q3ZAbgR.+tVL{okyy0庉~U׾/7ߖOKbp%g^_?Ԉoss *3Vi)b2~nI-}h:74JFu3J.~42`2"BW-<]eCX̫+ͤGtUk{Ňht:]ejp_#]fקt]e7tuB^!]YΉuD!\ڟu2J!nJ9Jstu`tu\K^Bk HvlJ@Wrcrc]Qn UF+y Q[]bU "JUl_*5tQZ=++.$}+ko*}ZuB^!] jZvB.<>9C 5n.*N4辩iu1Rw֒=-uwCi GcBM܃ Qy "$ybMWUO&ף,)$ yHJkW"6mi<o)桛&o"1MV/jyRqB׋PXZ\ʷX.SXݨF7lo XY.j/mk\;i RA(#`eNt)ifzR)bT@]`6Ei&"x(fA`$gk@!e;jdbo^Pcp[*PHV'KYSڳҽފٳݙ+OQxj#ĭ@b'n qȨb/+MЂcUTDlbB H6L玑4-cD3.ѝTyĄŠZ nEa]O,iel:2^u[199՛,N&jrJ2>LΛmXj#l)ڛpYJĥqvęĂ -a xD'!5TH"DGN̽B,@-C(& 3A?8$S ڕnX n>l L21v}{)8YaFd#+Gc)`!wr ,$>k;@A%@eqEJh^(SoR=(9ГfBsvaF,g7#^NG^խruiˋ0/^|JQM5Dhhm(F\7{dI1љO]bq,|x䷟P*~X>7k^|s&bzwBя(m{bIpKZ5R E^eZl[`͸^{,)W_.(WIjs_'Y;l-#>8/`Ȅ{1G;aʺ;.an?d4:uIb{ǤMQ΢Vv"9J&rlr&,YkrpBCgޗ<*% i!ĩ\kT*D1%b orgLԻPBY: 4=xs`Ke7=ax5_ݟ&է蠽_vX>Si?Eo0xp )ފC,ʽSr-8% NY[1a  ᨓZH[ lșr^',F.J TMBBC"\kDfZ##r"L"jO8C%&bpvG?0fk:O -p[UG<3&kvDxnoCo<^O;sAe!6yI{+D)0`ģ3IB >(`kIk@<"t~J=|oc#&Bs'/#􊲵Ϭ8I.1PV1) Sy(L}E_9zL^)T3iA͌ra*^x F?h}1X~>4-G !C@d.nxb~,d/ I_$|R &4O%Kl#3R24~TjyzXCtHZʶnck&p׆@]qM>Onz edu81r /0n톾tϏna[nsl}nd^7yirjѭ4:#zj9ep軳s eۈ9"/k! | )EOP-1Mn/8PXs4V|rRiSdNY^D%>|>K>qxc /EtqTk/ vkoh\5V@v}4qË\-xH}P-`)D-fZ̏!§zREɬjgl1\[?| :mQ'[cnv?ԡ@̓ݙEKlTKǗ7(ċwqb];{xxċIp5i`ݞxbHm2j j-;2V9q?mGl!&f&]둻{FKWa3Sve Jl ;7ܴonrmu4|>S4er ^*-pA "83^&)h Jз ;0E6g]:FW# }6+OGCCj6s, bNc"0Sw&* # d0!,!3__uhD18`Hva繪9JICa*TO_.>ChwF?gǵmu qEgW"׏od85&C{)**;]/`WR&!ɸfVN SoԀ[ƳE v8%S3JNH_:>TQr[ǿr]gOy|~a_H:fqv]Ot.q(oYB5s &eJ,˥)p+Dpq-@nUe30N%:W}-Zg[k%A &H&|T0.Ypy#,F!kuQ2ƚR LFfFܥTMqKI$j=ѲvRQ{XI& pUdH8Os17wl&NeM,LZRP8m}KeB8,9k7K zcx9XP^O1Gm#1ɜHBR b tٴ]d5rAЭb_f YHoHĹ4.39@: ~1im$2qo|t۹$xY d  l!mNhM DP )e C%"ג=g&ba}Ozh3Ԟ[3ay sSؾo Ǵgſif?lԂNɠ1 ׬Ye8\HD-} ^h>zչ|+&~~脊1BI뜏JI>2YsT ٘#s} nI5qz^Mm6Bxz@GUA!Gph@ x{g4fIQko 1p4ۍc;\YHWDKH?' DDr\k7X[#|+d<&7suy9f,ITBHأ2Dro0rKFzB% ,H $jHp:\GIpOZx, pWࢯ|ҾRܔ#L,^qtxa',^elCQ i2:AJY{vΉ3:섍ܘґ1ICU{!1:#83ʁSKQܲ9Tr4d|<;;_k̐x+t.mZ%6nخuoy:gR劵WACYP!~P@3`H(ָ+@]ҒނbvG}ڇj20-Hz$Vj %&2)Px6&H ހ UFX9i 灁2wm9jb5|Rkz @uHuƵ.JqA~רm pi f[yTtnSw1۹t%CJx'nNwF $>%A y%yR2&0v}7+Ɲ xDeV~g_r !m! ,Č%׏LBϽId`"h-Lfnϋ.>>G]SBuh37`;ju`U-Lt,ttm hWP*gc>*0>~$J:)M!)L6y2{h)b6}r"JY\r2YK:}ɧI vgeh]Сs|@!GN0-fg㈉s(["B6hR+++eڱB*UZ0AhFsiGB!)ehPwk qr>Qpn߬;eMޱ%OH4VZւ0MV z %>Ӫz<P=? )^;IǸWȣJE1ºhu`QN.#!=:Uf*~#QfAD( f̓FR 6#S*ICyS^g 'E/:=w!}4[ 3YZiCZ[Fs-6eG% ts4ni5O`a)o !Hm]~_ Ol"kn43-p )(g\2qp>j%%M̟^Lg$ҩϖŒ.0K*+I++r%0|X#(\FB>j1v1Ł\[OՂ~^Zo>: [|T|@&%+;ݲBv`5I.8#M>\2.`c<׋iY~?~bv:0I <1g(?Uګy]3 F jbO`'-{nDc7Ț(h4fb9ǣDO g[{l{A68V0i1(#_ǓP88|03F--* Gǃ0GX[?~|[p)%d2L T^ W)΄ߒ؇)N7 E,_@0&pBXLǨ$H'"`#KF$ʽ;v+$'n9<8IKەpnz >a,EA4>T 5.V'TN m: `fb},E;w޾DȉW.nE m؍[m<~/67|VgZSit-h5Q`U%8+HX? ̃k5'j6;zfjR%&ž$)=NEV0kCZI%]֥2S%aU0֐ CM涁Qk K.!!mx5%<+-j7d|]t$|r5~繠Oա`T*r/ա v:TrV_TRGhiZAN,U=@O6"s8^%J.4o,PdzH(XZ!CYh&gѡV0ep'mhO(-s~eF6'8k> Y9G? ,h <6w%b֭;fL)QC8ΫwX6IQ4 H>V`lNRAf" q} оhGk*CT^SVK,ӈ ڋ%A1LjZĘ+ڹh~$t!r_Դ*֕ 8O}'(Il}Us|y fL@թ.NjI6vD)@ T"x e8TCRw^zc -hVd%{F-o^\J 0xmJuD!@"3:s#wzt"ZcL\6w X&ulCThXMcݾ.5TOǻ|W7YU38n1{GӋSx糺MfO>g;8::wmozbDrxIvEo/oW7O9;R٬7+zZpn-_|w ׇZ^{s;?L--n_yn._a:F=am[pXs9嫻&hXzz)ss~zlz \mҡ[?nj`V뎉\}qL,uDԗ1:&Q=WD7⪐v_U𮋫BKWV#ZGV`F\rqU5 *5fĕqVGW"Xsyս5S$GRWW]znEGlވ+n$l_UVB%^\@qEUjGun p1Oj KRjf{0 j0,o^ h"h0Z!+f\$dD%LR䕋PHΒ !U~quY>^U@<](69m퍕k6ay 7%Hol177 v40ȫW~b3,k|͐5Q"Q܊j9#?ˏj.zs>%V) ZWkVk˾ U2g(9]![z8 gۯog?YhZm iDNP۹Xh]nlZ[[mӶn>iQOɟׅA~{ jTB XPi{%b5iUfĕ4FqUU/⊨. {qŕm#YE0GG d C?cʖ${wS$EӶ(ɑű]X9:u#4!sݥ5_+8Mn'AgE+ݛewl;|jMys`[Qp|28Y4ʼ: ☷gw_B` 4 #8{c]ͬy&,8>S,d1,vVKA ;^*O,'ODf7W ]o z-L}qڝ,V_R:S e?=_Za+3n ?2,?Y@ZZd҂|K Ć:--¥)]`7x*t%hӕDW~M6M66Oe]PL5ϼz&q2}jIEC:p|z>]rM7P~1kqtp3&V3܂ '׺sgo1+L=O[j5deL_uE xeA&xG%ְŠ\MlkZo9O7ůVkkoYf6Ӑنs|݃rƚI}MR<~?ߨg7$£ueӹXm{Ϛ/@ok:O:}C|8`jj~miߡ8 dmь˻q8ŗAEv87pǙp 9ўo9OC[*FدO[h2yrsf i=Fv;t^0?9@z8r㯝FgkYe!5N&FxKPD>Y[ajLރZoIN[PL.,M%Bِ{wp-V- #?EoG_2A?vZYB ӳ\_dRBVkQy;8Vz j뭫)1VeW]9ċs=%^Ӛe)- CK2_Hʏ W V!2+F kH u.Ds|CCd)}r˒wuhw 5>'$pmJ㓠}o|kKɧƧSލh:t%pd4m8 6%_#]7VOI] cXt%p ] Z>zJ'q* 4}8W^nzq pxv_x  J>. ;U8վׁ [&CWWTJڣ+ADWLF ѕvӡ+]>֪+AiDWlfBte+o'CW7j폝%]Bf*Qn*^X{Fim0Қhk7v2kvciA쉦_!MhLIU 'CW.RY;] ʵNtzGeϨt%&CWTJpt%(ךbNtz*Ĩ'fLG] \BW<]}tY&DWt֮ǿv%(5aJ Wr7BW/7NWUhx8xtK5/tkn(푭]*j_A]8h;At%h?vtHWVYJd p S+Ak^] JgOt銝RpF`'TƥN:L̛7SOw\Xi-{}^ɜfm9x2 ʻgĕ}:3-,;hg\TyS@o>JךZFeXդX5jn7Y\{q-QZdSY}a@P:}Zx oGUU%{gf۩͔hZ M 8ڠ̱ӴDӯP3+Qk ] ZCNWN5UШܧtuLz*t%h);] J:*jOvJŠf ]ӓ);CAN૤+⥧DWsgj5u%hѫ+AiՉ^]}u>IUEP;T9[˫#iX^_>&ݗw~y)䷋NoO+ZQY?=׷N:Ao;dn꺤A z/͏o70t5 ޲X_.J7\f5lc};Չ37!ǂuuo{:xg)w+>[,}+>XjE4>$Gk*̩l0<9U3ʧ'Ǵlt~^17 O͏)F:C\@9o/rs-6Of 兿{[2{,ZboG[d3*7ŻcҾd o |w%$?Bx緭K| Àogj.%ozE)W{MAeNΚ}v:)gI;6cfklrʤcrSSVی3.)՚nq`ԃ>}i3tZ=>ԭ/Rh]U8bܰ*O q0%:dՂ1'„hQqH~L<ԠQskV5Qèsbebje#?~K;Wj6և.ϵSv6CI]ӔR1 ײpT_z`1^Û1+176{PtPQ-r/9%C ^^;pfb21 C5-gU݃6\6Hsd,Tx2;LQ!C!si04T{!QwEF=bT{x!8aˬ+ЈcŴ'iOn7W/bhSH{:93Xsr>L\|P(=͍DUAɛJj=8s!+娺 :'ta۬+iunbi.If)ƮHHI?aA&ҙnLVCJ 6Z ЗBc5_qʪF'5/[jHQ!Օo*$66zSr c搃)F 0jA 1Ȏ ޞF Cm.5뎼`Fڨ/3ޅ Ҋ&ёgS BFnQxiW N7h A^Ak!N#4/uh;of![D)QnAWϾTd$XLytp ʲ/yXٹG$l K Ƀ>k +3\"5˶ CP{6U. )ٰ֞C*x@%P;wMBAQ bT6L){AØ`; ѕU] Q\#R}S:n3o%Cez|=rd2qm\ :؁YWBnTzC[ȸALAAX7[A( `-e3] R}\1<c&n$FY50MAC\JhSQ(%c@9T}H fNvu7ud&]{Xi TLgvJAQ"Ł.Hq+ufI7KPAw/uZ~ ?4ՕH!j{TQR@n$7u fyJL#HH/eJ>XԠF8(l,@3| C$F=(a&}A*p=N6 ԙ[ܠj3RUŲb։,4T U[PEIkBFK`ߧ9;À&/ǻv;eYvC-RZl/ LC6=2X c[oT^*cq&ʪd)H|1YZRU2d`SQD/CNF:/') >@E&rZ%d^Se( LǛRqd%&0A)`IQ'k#ԭ-2H#3xhS`QՅQ,TG7>/4"jdxnPMRuY+'Q O 06ٽ=+ȓBEǞX6'Kе 2"1wP:n 9(P"QPأB- c(#L϶ASQA>a5K q[8 -I)юn<Qy@>^!bq% ,hFՌ.!8X,pZz&/CQȢ td.d݆`m \gT*JJ/ P?AjDQ;SQõj`Qy8 16 c;gTWH!(Dj8{-.'h=VwD$P55Ԁʬ$޺;6Rvr魪ˌ*"ؾ,%ߨn$ʲky8VK `J. ڪ.lWWk.w;XC/]^h+Ε}E`0u0ugllfѓq۸*=#|wR%;YXv Ӛk)6$I#e056n(W 3Y6TO AĹ"* #B[OCA,WRsw[zP댂`@23POuzR ;ac_XB|?XbDtb]f Xv!?{WƑ OÍ/!=3^_-%)"C8|("BtW=]]OuFJXc%, cX 8#6lWKa-s yŵ!(`-I{1Z&q9'\ODVv %XzDI<`(Ke$ig҆'҃JApͽӴg0werXVA>Օ7Ѧ)9* `-!nCm#Q@8 ,xBg";++B?SpthkQ2(ƞ zG=H˒c.02pgAp! ]X-Su|rED00e| 0,u  ]O q18 fi DTWAC ( 2U+%)tUc.TY:EZZ9vK).]|vI 1_~Yz>HUY|%`NƂZ%zq?lZrޏÙK1KBrk JF}WU?~Lipcwifeh2ӛe҂|QV8dRZ&ݥq??,W}^.Tnm!DCܠXpګsץ@g\յ!b՗^ `Uh>%&tBq9&Hk3@ʥc~ @FüL d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&2 L d!@B&em)1't&+0lH$B_0O\G'C\OJd .Y$.:TkN@`#UWS"k"\Abƞ\ѓ".gWEZI -'Gs>ag?hƍ'E-b-Yaz_;s f'KK6a-? zu˕ɒ%fʑl+)>8/?qAug_gQ00m4~.K,xVC/Nu ryg~s]y6pMW*AS0sgccPUǝQ8-%kNxdަҗ„YۅR.5黳^bn`*gͪƻ:pr)zOp$A\mO&i-e>/s c8/yY***ԃ[V4z^,7͂qیu};4äjX1+;u=RU$Pb&2 URqJ !=DQè+3礭ˏCh ̚oxsvм*T(0 .@?LCƷ-:Ƥwߟ^n55pTx5./QN__-ĴwMA2k)/8*,yEU  aޝ-nUM5WH =xβ:22x̖y&Ę%J5Buf3td/{Ѽh(bi1{0UBľ4.^3kh֧4Tm&ܗ.UYsU.Bc4&4RTr׺HV:|)x]U{=C73җ>(4%~67+9] ў{\`kWצ\J+S=x> \ g"g٤OJKuNTԝ)Hc! ,Qk@$0NES)R 6VqbϕjiXAowtď *kK./2z'ZgغڥyO[g|TLn+<-svj|M(ȁ-{ts%o<*8A Q0tb:)J4{0?_^{M=kA9h4Akxr]y6E aU=|VmOh .G]\%RGn嶆4[48I6 \--0Bw59o-懒O1]߮u7 -Ŏj漭L/ZBYulܔaϖ(c&B]MXA&[{{pp{^K p35>mn|a<1'fM8}%l / &֝n0"D*rHs4ۊބ[S[f N R_&/ 6q$G3NK2SOI Q8\dzrL89l:Vsܕne7P ldSL\sDeAUp$ S2z^?U49sgqhu[}$m N$P*Td! ?fEۅjvc0Jx+}Aӻ$Trn#$[f벣;9 Yک]jKs$v]3qv-ݺ)n}*\'qH-۫ϴbR9qNlaV< zrk}JE<G>ÀuĒa cHPS-LB@Ι$ sM1Gm8DX2 :l>V3q5z07n,+* gHID.<(e[b$ڔDcy c=gz7n4ǽBY12 ƨV†]f4{.\3TL_'_90woFOY+9pu|/Sr\.= 30J sB*?م+&,t4x{tz4\Z&PI5D2J.lfc7Z0P]@},K}蘊1B뜏wmH4B0d'Y &,p؁-X4YwbwXl,nc5,z0`C&Wc#\\H}v&JzpYOޢhV+X`uFhk_Aֱ%~gI' &s4ǹ 3iʩΣт$<1V HF%vhXЕ"qI{ "Zh ^2ЛBK gu{\@f<]Ўu LJ>Z/n^۰sT2Ԝ.YMmJ[ pGֆJ=VA"RЏ.4LE2 )X! v3yaS۸]Es98=rp {qʛabw\*M5 ƤR vE.NWn/gq:9g.]0z 2<O1FbwO)8M#MSX2\jMZo*t7#@Fu{ ~뉲}ws;W-U` m\w-̉` uojאS] WL}n M@vTw0 ū;«pf2uq2 q1nn)+" 5 @"'ZDmQ&v+s`Ip=k1AdW 3O ::؀Ad,͍OB옜50ۢ 䊀9d9 JֻXq1xЗy(8 _ScPA/7Q%4oyr EQ\˾V@w X vY>*`c@@G LO:eXkAy;{f0l3e3B w{ⲉJXDTkvJNI7-|2Ƹ6\l5`-S%Ș)#j9a)'7ʄ 1a=36py 29BwIr.<Ş8b>Q0KMI7FPz$vie{Ik KnGM|;|3q@`%4 }w?TѺBS!4jkv}yzS}QBZ8Γ p%/PDKP$K9x.,X'=?D{qK}񊭇mbLwƘlި'0Hhӟ @ ]@xLI4P*TZ֪xO%U>HL 1#UJ'A1rYW Դ Bt6'HD# т~!*`ӕ ї?[g0 /bbJob^tI x4{yN12Fh&I(oeڰYlPRo!c)yb0FKn{\rRlt(NĐF᥊(Ek18oPy$4a*z F8 8zf\J6 ͷ$]\gāۇm]JMfmILkT8 \[w;k}0_Yg6@.~͝'D$mlh`3Z]uJ Sx퓠4esiT^]= go@2nǭͲB($ujW w5 ?\ymFs%WCNg>^Z/?z.[yS֜tRtܠKB#2ϩARsͨC&cNsku`'+EЂD=&E^8k'ArIXoSOĺzj|P'>hY~Ԝ=mBp&։k\4>'TʉN)6v#>/8=j]qdZj7e8(`crbHXh& RytЂ b\ ixβ9؊9^/a4lY$&JF`s*]50\x! VD)d"w{AccP9&Mʞ{T_z8a$$6̐A*Ea#7(陔,: [ n{ښYH3 @8PcQi^Y7gR]HˬWIXq#JP/怽$ iP"i{slLکC0xY5̯ŏq\LZ{7˱ddq=g NOInU,곜\7;ѧsF)BC w:w4sQvD2e0%0LBIv4׮{{<>]gZh96p1Fl*Kiͦ$7¸uyq G`Tb?u*ɇ7y`=10L_;\\Vl\G^͸O8֛l7rҖeХCl~)`(]Nz4ks蜨HmխN:ե}ZZ8YaoF"rL~u?R <{UYW=.|gp E}x~?}p=;q#0>;-#Im}Xg[9?2Gz]gsF۳ 5g~&K1.zU?,|7 ?؀jڛ44ie`vlKv3ס!/ FS9/ @x(_@>@-1:_ԏL\G _h%SNHLHr J>x i7r!X#=aDr|ۣ6I ϞxCLӘi*E^RmDD Jy%L)MF9ǵScw:O`C]Wsxb}0? xc6gDMnvPՆ:w: vUnv0n-@¤-Q˃vccL) QQH1V%&9DJfmf{S@-;_2jap(|7oP%P%gjxəJxɯ0^Da}@*,L:@lUV\Bq{a+ XR~0*]ejwqWP\s,#\͡+VOT*/F\13|Z\=`n#bԪ(R˽W /zb"K~0*F\ejym.JI+^9۩nƝn[N{1_7q9GK:j~/. {?ZAV;Kιh<t#X~Qlf:#tٚ0L[R+Ue`Ֆ!rky\ܴ~#@zP"p@F Ԋh(h楠lygKˎ_bMeQTpk0oE3OHz0a& uTl"D\0lE)(#8JFөkް S3˜$cLؐP6)xjՔ3>%dI># Ubgؙ\E**vRV~*N}_Ca[ peϿ˷32FLH?OKK-\FDQ!@'D 9ojj> uB <9ΒM; ^yX7?V_k1O%~VJYI?+g%~VJY5kp0tR˵$JYI?+g%~VJDPI?+gJYI?"UJYI?+g%~VJYI?+g%TJYI?kgHJYI?+g%~VJYI?+g%|QCR:Ym`^γ;1y&hi0cx#,q5$NyԨ,Q@!qgouFnILɔHFiJ)qj$tIPDRBHψ,14rObݵBnR-fyy@>W˖w>Z[g"X]q}ZnNR0u7=^餴ǻJw|o܉j.RKq8)XV]0SOU,?-zJb6*%J!*  BDNJx0USt qLv|SN{x`Kq1ʻ8ȹ8<slk\BlO_ Ƈ>SܮWx7oǓqquN 9v8pn ^XmV=Gwx}!9r35"VF7аB#R$pcHMN62Dpŕ&NhA)Q ,dRG ϒO!Z橏¡h%s*'Ӛ_XQMg_m8qzp1HY9?|<*)Z0)\-&1o2 L} : BE :p(QBHΑmA24=gIIIXTNEJaҖt[rJ1[XL2(¶-<-ܠq_eϸX`q@٧fѠz4_ IlFb#LҤQ?9I>*-$XQ. e}Ynxrg#9Y6y%A38C$&p.?U:t[l7/VtkiI}|o_'K4bNV8MQ1n\J'3wV@JЁI*mcIFgDQ x"3^3.$Ǡ:jk bo۠~ c_,b18"UX-* u?S x$D0f <8rH«dpFHgآ %Dit/TSHJSò G**y4c$̈́VSsq}aXl:ŝG|Y*W:IɱvQjw pRMu 2QD\Egs|IА R`Ma (%C]ysv٧<^kAlս?s ]HhHُO &Z \}GAHztҩ$IC5s\6$hd@'yD%_d;-zDJn_O[ur* 'yFHR΂V䘶"YqMBA& cWv xPJpB&a@].J h <1&b90so.EOGuc4tвvq.3[k\C?k!.[>}x>qUSdП4%Ϙ [kւ3XR0 +y'Yb\ C5 O`Vj KD#ub( QwwQiNQ !n7Iyx5㪏?ГU5Sbw v`,~:fWzTqMR[! 'H؄EF#11)QyXfҚ#] ϣ d:?tjy^1G^TNFb^CĨ({ AR<|yB+@tƢ81 8n|}M  a !:l92 $;!;#}?/sNAoy,]= rd}!iL瓶cN)|rybqh"?tr1xoUˏ2JwpRs_jMmFzxכ'vOC"vJy&-/wuMGKg5w$\=0B-reW79E!Itv+3nNX]ʯ9rdн^ɰ' ^.plBJ2ͅ )}=yƛ?Z8I$6G+z>ZNk":qV̋2v^ ,<5s.+glPlg4~|QE)iS\T6Ц/m "r-9\a!Ah#V~Sm JgA-ff65?5Tk "%-o5_0di*O˓av񏞂O%oUt6\+z9Sή ;^*0s%&x[ m)J5`xxxxVx!(F`$8/zC4J5N[;9ō!APic< R#e18sUDl ha[0 ^И_fC U&!Ϯ~+޽j< )`S{_8{;v{NX%1{L5G|uPol~QreQjWr~>X,bR%cUq )&=b!;G'gxT:ؒ=7K' ) Rx±iR5BP0vY4a#^ wE%Kfjyx rm}cei2bhI ykuEޗ\lTp+7VSM.=;z&v$d_bKX6ۜ &c0I^ĖR2|%gS}҄ޓǐ$v*S6cQ1, CwrS$r"`  m]8!bxC( 79m~6FGіd=wa 3̬vܵZ=AR+ڹ(S; *fJ >0 \Ir3)#eP˭X[2NyiP%QeAލO8+ŎPմ ?-֣y[6y˃JJR?iBJ -/o)&YH ..p9)n0Zsҕb,+ܹ xb%͙ iͣA#hT1I c$1TRD=ՔMdo 5U1O4Zbw/׳@&^UU^?|;1Ӌ?aǿ8>M&j1uZV*öO(C`n~.=t)LjCt(*ԙd-mCJ*{:Rgyρ`5QYZ-Jm|,SĿduh;lH_uW!u^î$1N:&42h40(/׉z7(dY3i >ii'ɇ`eKC<.јgY/< 6ɇD(N$p-%#<#CJssͥW.UC>HDDS\&:˚-.C0!B*Caj3jX$"﵌ &Z'%"sx6s(ۤ%˕Mt'?\#'j6!D >2J#U!8> , )cJfHjX H6xT>s>`.0 DpK{+%R.Tr.sOp>rIie2׏UimrLdyhﻙ-O;M>W҅$AǷ]Y"Mwj~ͱ=gJ{bhoϸ+7YQyMRmQ8cF2y$#`8aHz,"+2 %ZE@'DoW12S`A Ɯ qlXmd@Pf,= k3ޢmTJ9}kd7Aij?=-bK͙ldE(a~R"!Z]q8 0S emHΞQk % 69`^e'h]Tff~Ķ_q1ۂFDZs|=j vdCiq05L)8D&QR-GQg 1:8+F.sCfG9AQNq/{N09rZ ۂDZH2#"G]#^+u+\k.9"SoH ^2'C֤9bU!!^: )tKb-DXҠO0IXُ7\j^gQr,.̸H{\qq׈gR%3JJMAG`V9F"Oy.7Sb[08ef<< l,\UkNosyu[JG?>OECiq`Qd.#]IKk0%+W49_7=m,h? ~1Gw%;.aiNޱ{?&=( KHâ!pY A}ޡZP-)✡="Xnq =0<00 !HTȰL;6ɥ1H=^N;Yo{،-1pL?U n*Oob-V& Rb[A&[ YpFB$tr!D Je+DGx2=ly6Q 2FPI+Jlhqw{%~'W}w42t8^?3rt~0JZ!bIF;nR6QXj` LЀ Vőg}YO͓uZl9C"C:.~a/4FO} }#Z=GzN%z=#j ѯeZWMg*B~NK7KqMA/ƾ}~4}]uxQXڣңU(4>CjOF{_|7'/Q:rW:J_Hm n"k}y:iZQ|a;z=ex2d<.~ ߧO6cutFXi}`<\g< ~~0ߏ߽ہJˠe%1]*t*,Bk/~cQuus('^E20_͛wsi˔Xji >x ].{z;JH[-=K.=MR\ð9o"WMr881sQo3JΤ6xs`\5-cm|@ SU>>;I`L Ug=YyBmH,ʵ]Mv.#pȻc{L,c05IQE57V"FMsFV;|qI:>[>NׁY?Eh{\$/l׬zzہ6{+"_*j]4gv"3#̥3M0FW攢ID[4-)l{-۽\ό78G+$Ng%:c_.U`,Nu7A'_n "$2IV߷zOY(j( `[4_TWWlNFݘ8]  kWXV&U ai>h G2 \g/QMM$nc - 7%M, Yrδ6w6'0&`"P!S˂ FuW`)8mm֤ykjiNg(õ[(` sdl޷3\S'jRo}>5~S:%DFOa@{Ҹuqu[L;|踊1BI뜏J7tg~d3n4 %h<#jq<xьŶ]wЪB]OΦP?Wu?e !g"dĉx{)T(u"' G5M2k/>XJ1eC$Pd! ?' (DrLkd A M1tbĹ߉^: r2vkCRx7lY-%?T1"K 9/{#г~/ߕ<{X@yrpù}>L@Cq+31@ 6o Vrruh mSNd [A)" {qY%ފ=%dmv'Զ.\ kQ G pAO VnHzR20v $/ʻ/|txpuޯ ,\fuw%3tE>L6p* ,NG$`s.gijwU2jE߆NM!H艔*7QIXNFIޢ :XZdÜ(Dž"h'sJDɐu.%)/Rusm|`&[3,+ot׸<(4Ĺ!8S-d gpE/>ElѤp_W܊򥹎έm{z4#T \ĵԆ)0P͘Q)B:(}LT;}t^[EMK8H2fa,IԶZs$mZQ[(<K\2L+`f1ܺhuQ '*Z*n!5v (2R( f͂RdyR!UR ԑl@wySw02_W.8#Mz涖9sQ,d'81gvtJV3-ݫ2QV?y_=zr620Vyi\ABy8WT1;bam_߼ԂRseAWg8w?pʸ9}N?waLR(DC`oդ̞1q/i`5Q*D:wovfgVcO({VZ -kÅXV& !p*2S+@sR ƅu)Ҍ|ķY2eVRB40ϬA &r1qFP&ml8̼+͒ \;>;WZM>}ӽAË43V^zGcI x)s`x>6qU3'lc$1 dɽNCC/#Bډum!DmtII ֤B(:%b"J,K(c53E4R@XkȱȨkg ߸>b= "Z}Tfk,čv۲YkzѬDq0ʋ1o"hc O>i 6'JTYp\6U:rj),B14-BL1&+܆RcR_uߞ 'NEUAgX.(;UXNvJ\{(٩F=;Ueze٩PL>>!O7Ӓ'IdI6h)]Vˎ 떳a:6[Ću+ذs,P2ZEL\蒗b9#hf5ʆuJ.%ja6%V(LD!GY ij 7&Ζwzk&w럌nknnTsnu6x6jkxWo~7I$E{/W,}.W~vNm-Fp˜޶f|b %] _Jcwt^P}7bI%tnM͛l楒h8G6o3mobbܚ.@˞;(]j9k趁>܉_yJs839 slNjbQ\fn)yL\0iQ```LE\ŔXvSbRΔ MB-"9"ePZCMUN]Fue/հn'Ƨ:vT 1+9ѐdg8[2z=[%GT|v > -kOԘm߰i Չ: ʰUX^INQ;7>_]灳=]SEB.|珟~xT*NIH$I uӤBHN  u|dWRL Oʉ.fR"øR?jdnd;v ۨI)h$V$ Y.*⪃i k*Rڎ FueAZtP}QPz2XdЩĖC/(hn ,yfu0qz^u0i3IdhPWSW}I+XS~0ꪈ졨"URB^)H]ᨫJ܃QWEZ mWW(SWP] ]oFWpw~?w{I6{$;`3cȒ'~Uò-YL1ѐ쪯Etl ]\Et[ 2tBHWBad]`)uk XWT[WCiuGW$TFӂ}x3& _5UMsv*F6^#M+- e-+|(մ-thM]BZVYb@qm Qvի+-%mZb@Jn{VDmWD%#WHW0 R+۳"pO_ -%_E[|J<1ŧ*+JyZEWC)"**Еخ2ZDWX5-thNfJO~]55a1V4]B2DHmZDW"hk XWV6eӎЕ%Ӫs (uЕ|b3B kU5'B\sZ*DtU h-]ɎzJl@|5ڴm+@N0_ %]BbT &+* *tBvt Sk$k]`KDk RB6e J:˰.;k;h 7@4$bDٲ:eIee u42qpn_xtA+A,.k_@KЩe7,sY {q,{G셯ւ5[FP~Oq_}7o[xp\ W/x =TM5'^O؛J}6lttֻ/zQ+q~)Ho,v\kn?_%]Y,m r] E)"qW x?gIO;Ǎ4ᣱ&B'- ڨ<QJOuLy7S @nƷ.kh\\h0u뼼-_mya߻}7%We6-x፛_l*9rsg)x<;4]KBXσOn2(tz~~G凙O*}W*ݛU^mۻl=>[r{ߍ@FnMzp){!8 /㏽ٺN 0@ߜmźoGqe]EAEd??LƋ̰p]UvD|$}wzUFCf4] oX)Xץ˔g\ﳌy'`-}Fx#[0ƵmcTq OҦ;qɱciC.ֵ;=)+\\ bYqņEͶBY8i'k;jQFHPG\h %Di0/TSeyG-'<fBsw͌XÌx}=X)W:kcyQ̋ŎxpQڣWb+Y*KG.#K$A^p$BF0B}p^.U=7ty\Zt<t܋[l֎r"$dϴ1'!\(Yd-#Re'XǾe&Q)ML EʜysrT*Ȓ371B^*/#)m}fA')RjɡcZu(mQ8?1s^$R9RFeǩfPgt Ɍrar2T(x9d$צQֻ}Q!P=?  &{G^zX(f1?l0gcJ~*ʈsZFٔX9w4^ƏJꪫ!4f"F+̞.-VR/? \q/wu[LgoƱwۋq4;/m<("_c(]}vy\}3-rap5Be/z E!X5{X4b[fiS9\IrY-Jx:x10_׬ GQk~y?Ggi0V Ƿ~;+jNVAkZhb bW Z}2 ݂Fw/Ɨ.L{eޗ_h,Ծ<† 9hg i4  n%F"wB*߷0|/zlZ~zU6^ZF󞃐?@- }&ՏiO*+pv@АD#wAof[c;qY&e#2w+NBp*A쎡TAydfRޅ8#ު%uz!2F˾}J퍁sn m`wW/P2hM&ƌ:Gf7$xT0+5RS$$SҢHJ2NIR].KR>%/u< 2ۈ>}*%03.;ŲCEv")(vBulEwuq{.:'2x̖E)xPLCw3IŹ1=3h;D.UV]DҪQ,u}+Q“XC! % :U0NW%7Mbٛ&ƝIysN< ~>'YvHة`\pI>=FN7?"rO8QDaRkRȅ'8wϹL*&& `-uZ'z6mP3LoE !FX2/!q,ַoNK'% 5ZRpP}툔` cHPSRnCΙ$ )樍u7$ .A >Ow NU){݌Bžre Xʃ09kD"QLw?_9/M3~ rIv'M$/~lb̐s_ 6ciytuwuuUuկr2p9G56e(2Ųłvh X,Whv;"x Yq̐%L+iC lsS|R&" R^PqiTu/Aԁ%[jg&u BHݏ(hÕ[6CQ=;^1Jƺ-plDM<[K~ADА iJd< )}mW;ۮwHZ1C*D) )pCw6W&P >fJP0@iZ ]yB(j^@4uﷻ(ϛ$m!U]|ϗ 28TǙ6(Kk/3 eYA񨵷uXf {U2uh+ %Yh$"95Z A bJs Szx4N*ޮc6ESl][/dY ܲE1ۓdtG&Wpzށ i27@H=I{)ꆛ{rK&$ )w΄2!a'LH겴lCBۑ3m%4G-ʊ=XVl$xRB{EnOmvFfLix]dl`湀謏IM5SegB7w]@+r6.8sry5U zxy f' npʳW3.X^]V`E%W_J_% qUsJRj G!͈<]w\n'8GQF7O#`4}3J(=\8?ݓoi6Ө\Y>p]ԔYRyܘp-uVK[kf7_YuE?&>Z⵻Vh{ٰ|34k3Pֲa!9aSU:Ƞj^~^FȜ60ժg&ף@uUZnhbkގ™a?hFhQժ+?]6GeeeRk hA<ע!f8 ؊!C*RTrM56PU(a8ɬ쁴V3`bG[sf#1[aFLr.*bZՋ7(E;NUY`//p2Fv:cO$`;|q=~'{-"RG<`KIIԔ#c Q]h Hr (1*}4DcYQ28lڨ"Qтg:&;;#gL1{5XڬFb4%2w`\]Vne|nZJnִxN83 pN%2H2]]t$٨VI)h$V$A$=EI5{0UKig{<; K>ִR[q&] GC۳iW߮/b-}S<|+g1չHU4r-ۇh_B]vnpmcPY茡N>GZFԩ[iRJPݹJufϭT hck\@rZ'J]#[}[zujKl!;*vUD<`sާQV҂Lu)g2v $j_)]:k卻:?Ӻ^WVY̾@ʁ=ϗ(Cd2O"3H6ϹyϒEDB3n2nuq=F͑Z-LCNuN]9^dۮ/djBD]Wԩ"[,r|$AlN 9J:lS\99RXMTw,1yW`ֺ  .#RR.ˌ40ڷ1!s5Zi*r׊mgܯآ&KDkMȘV@`^}T/dy~ײѫ/bZn^xvZ>$De$.+e^xv=ݎٹ9z HY*5J:\^)K^:A3SRRRfS^b18J`ȑ9TzpB=k M@{pFՐW7ˢ5Y[x] g)6+0nƓtI ",JpvëINC^ۘȵ DN\Qֿ܌ rVG_&_9=뭕3~4$-nu՛76yp7/ܹ $MG yfpͣETlVT3[74=iSρ(8/hYES$WYh/QiH)D =RERB#HȨF%M`/ !}iUqj&[M6y"{\h)ڌR'Cֹ H}mJCΤi_y we'WL5n:*'7g`q곅Ll2QT:R80ST;M cŭ_ubvk۞Ρ-k rp]'&a%k4cWD-PVӝ 뱩D9J'iD݋u;` 'I,%VksMV 1YEWEؿ/ĺsUpy,ڬTt2FU[4*WU 5=UDBzLE$b,hbQ#Q!6bރiU}i域S78j}9b8?gsVu ؙ<`┡Q&(?'Z D8pq7 ˃taI{mQ49)J JwZW6nr9zFnq; -q4SCSC~;_p&)U$Xd9_͜0 [oqSj|Z\68QRWqҏ p9 + 0,_ ߏ~tF7சM #oQe[6ؠ]vv]eE῕L ꃗJx߿ yr*҂/k{XVH|YHGDf" Khy9(3gLP=+rv Άp!s.f8nG FIxP8!\:3Q l8)_1V3q\_=}WBs=O,^P0>ds揌|pXs%`y.^N*)T/9 _Pm_ V09[}Q)ybp ;Kbxق2( F824 cTB҈ BF -C8hP3|+N͗ϝ23"̝'I*w(g(jNsl:&MYJ`Q8<ߛYbo{w cnդpϘ 4"bX5Q*D:wWzi{V=Y冞Yӳ`K YaOjrSU,$ Wց q ݖBl!ڎ1rk28ާ{> fB{T0H ;-wxN:4pbKKN:"<'c$1 dɽN׆^G=;>tIR+7сVPEYQƄk:g2hLhc_c5Wr+6=<(^ddT9lmѭ߫z/Y{tTtojrj#RS#ԫQ{,(ȕeè2aT.X;!ĕc $#qU6#jr+SPD/ޠe{$`vD⪐k侈+VQP{-+!N̶:W\&E\j|*T Ջ(3좁A/@5 5-6pVX3w2'gaWg0)Ԃz+2H2]]`?Q|#ǮEXqd_ A}X!yȾ$~ʖ#ɳs-˚{G] hUW"2HL7O*㡫_%N&&\DOW׵C/:]8Y~a {Mǻ"J`WDe[YZ]M&\k} (=Е İ"`+q5t5ը+u&:]M)]!]쏮&Մh-t&Jtut}J+jjq-t&䕮<Ŀ=|B}+(/ߡS4f(kWf7Ѿ~P#Mhɯ ׬fnutt5Q޻ϱЕDUMa=kvK?`7/ăPz HW11kZ7jDkjw'㡫Кj^pjFÇNW#>_ڽWW;5/vCk_h{n(=,;UTzЛŠj~|7Z hKm eLJWGHWa'u5:h=:]M!*]!]菝"vj[Uzg5zza ,WSbORbO/ %{-++fVCW.zD+jLtut吩'뙡3/B" ԧ;PL]'\p_⻡{_fv(UU%M̊ =&\h:]e tut7WDWp\. w=֚3QKwdZ]0\bpׇ7]M/1LJO􊬘wl奋~ZNhyUځճ>E&"Jz h9xtu t_1H6~ЀQ:9/.fD۽#i]M}{{ɛ?ߛ7onbvsv A9jFm{ e=>,!/?\.Fހ~%z7?E@ ._wG4g-g|ZtO&cܞw o)6Y>iﻝpwF[̔;>ٶ$Pۼ\FmT㹷6gFE<+.9{GHDzo%WZ`$܍E3Zah&ɗ ņERrk w>Z1Xdx2B;  wù0'Z[B,=Ɍahq1ZcÕ+F-9S&a h`v^mɦ ڥbLiFCLK'9dlfS 11ИU6FQӐ+X $IQ"oHh'JL~o>\ds5zSAd*هX O|"RYr!=7!9ep-܇xosԑJj6 MåH#HdKF^֤O>)d b'ؓ9fqu$~b滌ڄ[I~`) VCJQ[j!rX=n}J"X'k^Vh*- $'b0#'-UC9;0f"QD!Ek2"XdGBnPi٥3P4R![=`}e)-G`GPQCљm> y9\xi@חG9kD(Q\zPyUJh+a$8tyqpl5ȥPXLčd+LŅĕٙ4ƺ П^v=VFU uڋ$l@6lmB9ԠmEX6[ V G.yA)>x X7G#\-"0+wolPQ&t$!D&, |AťEN& o@e5HQcیdָ~A Av% @(1r3kiAe ¶o}04|hPѾ4 6a̾Z X-[Kƒ]9  {`&CC\tYjM3B`@y܂dT  T rl 34L~a+AtbE,)tiU BAx=$ (Ez@?P֙[ ()bTW P\.ޫ( E鵅8bC(,dPg >=j_nݢłFi NHNoVU9u:1( / }4`wN~u:5O2ğk;U0\h3_uSw Dm3bka&a=Cφ /- tlsU2G$]͡YU&:cxꘆg>aIzA{ .B` ) >@(E&rZcd^[1P>X,.-ezOG3Ht_5\uuv<o Q:^z,:٫P:mOU;OdQB4D2X ݏOˋ }kͼȓSK5q:'Xo.BfȈh2"A]%6'{sP6D")B-TpGQu $$lLCQXfYۚ$gT ]\DS;6@D@zxWijw ,O߫f c%52 {dN2{=-z5.PCXm(utqHZm;EAi}kKu2x\ӄ XTdByӊPp5o$豈k5c4F*ygAy-#b rL0H5+[F?_˛Vv\`ɜfr{.y>PˠwיaɇӔj9ef2J a17;??N̽8ɟo%veANn7i|q1wt=6绡?N[s~`@ɿl||>{XyE2g UaP&59&WDi-NֆCwMN tN TJ@R': N uH@R': N uH@R': N uH@R': N uH@R': N uH@R': N Ĉ:&'3 ޮ  ijS'1:GNS': N uH@R': N uH@R': N uH@R': N uH@R': N uH@R': N ux@B$z&!YHF8ƃw\P'1:"KR': N uH@R': N uH@R': N uH@R': N uH@R': N uH@R': N usVL fZp= A ":B'I@R': N uH@R': N uH@R': N uH@R': N uH@R'wmiWk<ݴmr?l]$g"$;.RLFk] M$|y;|@i$P F@i$P F@i$P 4@S5kffGgϺ^ﻝ]y~Y\̖'dDijZ%(>p )K0pNs "[ht,qH8!J@37"tn q](BZ%BRt5@]IJIWL#qEWHkxJp@u3;[w0(C9wMbXqu!˙ȈK>&< e߽y͛jSWs )aSJǂ1| 47~y+*i*Ɇ)mZv|T <ۗXt_^Rڤ { NQčҪu&b,Ɣ]!2 qM4oD-4\IoD?ٮf9+~V8ꉫ'U/JK JWlնIW)zʙ<|vXhtĢ+UD+&jbGF+~W']! )5O8t& r"bu>IWҕDr2m;سE1se 37 rd~ʚ*lf[7)>̚rR-?N}1_U>8 N )|Zr^T:/(E3U| u%?C雖\W ݬoUU89?^z:c{zGiIkY.8qR1;E`+Gļ93Å] ,gV~3d<a lT^dֹEEl^BڛTvF;ggYS*KS8eyH>˙*ӼpȽSUUvWMUnަdq;mksK&Fpu 9ꨞ"4WXHk֦&$LCyF1#v)4!eLؒh4 Qki ]HɒiX91 eBJ!r1`Mt${hjOJ:QWV*"cj 0t:F\#c229yvݲ\}}(]6~vhZ FW(0l+tܢTHWFW+D,BZCRvIWS\h]uhtǢ+4t]!Iuŕ4G+^'.W i ]WHIu%14&]!4 qEWHkϮ@KJrn;C\1c)F ʼne=[SEEDF`+54M&b@fK~M+-(+WXtuIWԕ0pD`54Fue Q5 X qU4oD+IWԕ5ZYXx]!6 hMc2IW;mYlD,e _`Ӻ+ȁh@b=tŒ[j^/n`_B\cJ2fIW/+]!g@fA h`'-}ul棿7p ^e\zzc89ʾy?~h _\mAv>퓣&Zn;<6cx mI[r3o2{`wV?_-{'rt|/֕dnSt_y؟&ZpGmkGm)T@moeuio^'g%=dx)zLȘ (a4s *Ppwƭu7'_e-e߯ѩi[n׶;m_um]Վ|UѶ%8!ԶD`ˣIWQK\!d)Y`ƭd-FѬIĔe~%Lx0mjaiAULmb\W!sR!$UQ.} *&6[M[AZ%OM!6%фtE+ĕ<DZR뤫! 2hjz,ctBÑGoފXhDF`4x:"^LgwM#IԴpؒxtĢ+6t]!I+ _b} qEWH+6j8D+"W ?( M?EG1?o?h퇫F=г^+tܢLV:]! h% )JvHW,.5 i ]WH)Sv5D]qN -% q )M4  qEWHXB:2$]$pcaԎ7JR y{lL҈4 qEHk4P QJ`_Èt\ǣ+$W H{iQF uPE+\E+=~& )I0*fle4\Mi " )uzf7D]YI `d՟GWbˢ#" G/`M=x?\v`]ۏR]IW-z*$"] q9EWH+MBJ+S1eW5FWLq#SPRw޶&] GW\icTD`KE4B\&bj~B&] QWVbYHlD{^V`b\ OJP{[T>=k.:)6$ӀWĩX. kcʂ:*yx(tP\ w~ДzO~'G?cn䨭'G p弙Nۯ>Mٛy89[Q}qM7u1k>꽾6r?B~Ԋnb o"󛢔%\zSJ\VFm.%Rxѽ mSY&^WDڴtV< SEf>!hV وVzx0MS֙}"i mBfxDOwsĕѼGZB HXNxkdrԈ ˨O 5SqEH˃4R6PhZiMxL xJ<]!mY%R!Jk|XtVuJ'] PWgIW,i " i]W@)MQpxt ;R6L[c5=z~h?Z}Qې=t%[BcD+>BpEWHUBֵNz]1 _b} qEWH+uBJe+A(1eW̨FWl,BZ|vR'] RWΝ-nX%DmFZ^;zC r7Pڊr& N]. ݩEko)Ϧ p٧S?˖Rwxne  2/l>ΚOnO]ЫgUiƋE,קs{_\olgxn8?N/1vlG勪>]_Or =~No_fՅY/Ku1_L.OݯHruxj?Ztm=7Z8G 3bMܹwDt'_6Ǭ]ﳝjHW \P]rq#N2=":Zc`>\,_S-.|R_uϒqU;8~g,z%{ 3MJ%7E g5DwT lJS,YQdl5#6Z;Nnl2/ lmxrH?_|䘡|n&mwY=e[e'4w=z_CvX8]J[ۼM9{f)^(#E7ߤ<;J;ƽ11{EdAPz(%nr$onQWPN0Ƽ5ig*Wr^JH,ȶ+A$j !4zJAy骢,-SV笮)Ѣ. Z.z4c=0ΝtsH磭/'Z6YIyv6iG? fJPwau\WrϤuh?o"/0u V]Ԭ* ejK(H+Gt-f IjюZW3*k$8s22(@=IږޙTNE/(AR6pl*{s"Y.Dم6pC綼a՗cMOvqA^Lͧd[[)L MO Y&i~r"j}]2Yy+)sBY/ :krk{c% QD: @XUW&O{f:MPx>_v`T$;ŖS1i 21J<[9 ^u]jGZ*]%B2#+5yR)晩$((]-QSه{F}s(PxٞȒxJJ;c)TZ+uE$jmV]Y(⴨TmYXGJҐ^(/Vr8,3+^yE5%̈́慦‰lĽ|وN=u ^\ҳνEsETCcT{'kZUri!kʗƕwm#YfR+ Xv4v'tf>A=me#w")YrLI)[vB[TGލ$hǂ4XCciǾPwl9 [ z:u5Yl_6ָ-{Wa|r-J+=F+Z<7k7Ys'k6}|EC[tzS:=.~3_;lA-33BҖr"'Ǵ2+8YKH\KoT us<(%i!S㐐'S2P( HM cm"1kL Υֽb| \ oLQ/gv97x, reP{=#yp>|> ! '1!4% Ҵ\{Iǒ!E8EȺܨqozAXjDD0R'rpIx ) Qq5 ;5 T#&Q>c"t@hpp(L"jGs%DZtl:gaz p892E$O$wwwoɣ]"<[6pS/-3R./`!$p)MXd!I{OW1e&<'ЛV[fy hPK`$5DdE3p'`!x|oy:<[?zr+3Ko2oru<0+SF/սXieG$ nK6(ӏy|5?M1 Fhg_p8Ka>@Q:.ִӥJ+nwYcKܕsB c_:{fZJ|~oaNڧQRc(h[׀owޠtQ4Tfy4l$U{ϘLTe$D%A$ b4e 餬V%J37C> £U8QʡwOI(q:2w6^G%y5 (Pd&HKl-ʨ$Dñ>D,5 (9T DdV%ڻ P*n=\LC.9҄Xގ\Հ鳋%0ݮus\(Ku6&0xHrf4:M *8ք.nK5 1[ r%'&@S'0Z)L\ɛUχSpKRU3J*3ĝTI9I*ER~<"uv~:BrZe7A2cJFT ! vTGg?"pGxmJ051RHB텳ǹ*D&A ti @BP;gGG++HY3FP2-bqz{K1띂I!k qpDpr@FJBEGI<3z֚(qo 1 Xd(L(DJr^L];zzk7>X(>h;I1H2 ƨV?)Jפ45s"#CńQ=U)xY9hԕi:Q[(ذ=pEI7):y]Oh gM]窆W񂁤8 1vBX WtHY=5b'캫5Χ`^ߗ`~;ib:0s]9]+vu1S~j f я'U>% DGVU}N2O(r?ŋakrpߍq؟G[>O}&-Jl}H93d*'⭮RqJY5GoثAM _}6hμ..xQEL3OF~*I^.dG2?jcq=O"vyR4`?^HKm,Eyr6MTޖKyzISeW]ə! 3+j֐G![C<5bNR]P%LWЀ'$мt7ZN<'S',@2F<5Tq, ybVd$Sd3qvXcݕɮzӛ}4B.Bvߔyz =bXlN bzzD ,o(qRR(k+kL)b-.jNYb@#YP{%foZ':L F\F+yTK ژ`2J;#Nb!k1z"\0@ jkԂ'"&OFuu :gEKߥ<6h^sGZ.Ժ#=lA_jɱGGO2&ЫWu lF+[k^rPȁ‹'q]$swhIdA@M˛o?ཱྀs0TU "^ep-JE9\tB]Xd$(1 a#0aeš2;iKq.g8kʹ;d)yiYB[Uw-wE/o_NMi@4j"su%\/4˰ Em|ȕ^i{0:F-Y/ҳ%/?甅a uKZY_ 2ȍSi%JJ9-0Lo2RJIpK2sc珩ZNo9,ĹJ! 1htJTă@B6Fz@ժq/qIgEGxֿ_g1r 汭f=ST19 i!3>.rlodхR%HJyƤq\׻b8CSJ>y֎%gi^'iAlNQ2G8s1${YKk&3⩣`9i O]u&=e嚅~!yx~na{)W{߽GCq[=|"cC!ϊu/Ղp/)/#2Bؤ(=s@4 1(iS "QJ`{]rJQlt3$ Sxc1*! 1'[ )KaN@-^SЙ8;ki뽀zGڜYȍMsXOUâm:znQ )cv'Ul]bsn{|Nm64e-F[wv;o\wyf}<[ݼsϻxvr]uKM6ʛL1yX `E]+δiܼo<&?7SzSpwa cY;*+!`- B2gs+}B,0(ˎ.궥Bҙdu2p 3zõ%dXI\>RxuWVwiMSp3iy-k=9N:*>*P78,2DqXTwd\7gy\I!/gv&x5LfI؝5F,ב)ЗCu)]\ $uz^/Z͂qtthmQE (E mŻo K-5N>qBG_Qe\ ޕ[f͋k]TϞR?|z2ͪbvZyaĜa?弮8Gzi}? rbiZ0ԓƞ\iS7؍dZew$1 G5> Y,h8XLiz$ zm0Iqާ|%?FP^}>IQ87yK4TO1{Y8{ bǿ}/o˿~`ZnLo~tE$mmuM-F#3lA\9~q_%XCbđfݠ>xɤMޖrk#UV6} Lٜa˨χ|u Qg QTʢPwU}`pw2+ZTkξy.Fa{TH&po rK!B-]/[+ *P ϦeTn@0ƕ I N J8+RD͢9j:Fp4dyv^1MBx'̼ž17fW!\H92N#J҂!S.hH3gT&Ȅׯ+*DU)VYsUnqӭl+D.WYD xȕV^z[ey0I<J(!Dt|6^qRӲ-F-y5A%j)` 8,'ļ\ǿ~}Wŋ״8M7!Uo%SW.j[SI!}hNZ8\ /d[N=U?h nΊ".l9 3i4_e]"ZMk.CRd\޼}J71}OOzzkQ2[m,޻|!h.* |pFĽs+AXfgv-xkb&b "dP07&o+^8u5^}VԀ0:Z+WRG/'i؛Rr/o$(H߽^figkpePOоl&#eq)pJ>_\z%'ZЃE˭޴/,5[vxΒVJ펴b+D77.m=CdT=!(/2뾝Bewev$~Y~g l;>@n($Kmו %ryr$3 |ZZ:K^"yL߲B[\mQd[Jw\%;N9$S)csVhB̘QL2HZ]1jlKI oM)z/b:s"EGƪTY[#gJŷB.6#SUk)jR_\uY f4؜ȨbN^0 f^h9Qs jL[`VyH If巘%,W^F53eAm5})ɊNn\jf]J9dr̲rRo[1q:8@=-?6[?MMPk}**]1Gr&捴U;L22 Bj\H&$LGd[*2*s{VE*cH@ГBQtCt)#T\\ FȹUZ8c+,$EG%Eel7T=eV?I&͓?kP`eПLsV LI@I>,udH Ir)ǢS6 Bhm,̂{.C&iGA D`#Qhh*Smfjۏqn;Em&jo_fC헢g2Or/J541٭Rzɢ9oU`NzξXFУ^e*4}r,JFRB}p^&܋ zܒG z\vSGc_{`=)-9iTԭפ)c5" XnքH$D$U&3ϵRHs)XZȹ>#}!(VeI?OH'wjɎE[;)L=2 ?Wxw<$,f[a@6̓12HIB`>80yEA3OOb=!<3t|BK)mYEE-!N1!&ߘ:iy k :ό{/ ThQ! w6HW.CQΓUC#b9\^{`PX[gSR1bV?~/Sng޵q$Bۻ_EػKfMr\bK̇J~3$EeILfW8g1'@if*#5GՕBa< 0~5U?G xd_=rv\+P۸A/p^ϊoe&۴ _'u_/F?ex2{d<cqx.~]>MuϥGK괹1 F*i7Ҍ~3ojn$4U0WxaPK- 1Z~=q*˜1Fgiſtsf:57Yz5-0#*"_8 C1} MC4_'}_i iU?o;nK+UT~$uOzroa۟On!ͬVdž%t_{5/88j⦧nL',in8AqQwͱYF{ڭ >hJ;nrr Qh-cFQr6sR>EP218WgGA7ܩ]rBm D*dhL*ܠIۍ)PU&g҇J! .pU9d$a;i8J\OȽhi:3؏Zܽ6H&|t.Ypy+,FU6Q2-RGE,u5-^ɴcK T^4vPYgtL ŮvFΎj5=aLy JH5ڕy> EX޾; hJfTxNAfC\# (}2ZGB$LNĀLg4*ۮvF}jz5 %, K` Uҥj JЪJCƜġRWj_'X*RѶDKMݤdOO q-d8~zj3[3a s0pJ2p8f>LqG9H*B%k+ Ģ#zsٍ`IaK&YBdO\+g4XvcᣏLdtF*mj>}HKBCEk^?4IA ^hAH>@pB9GBC<*xT;U_ПՂkū uJ)]+ko|ӀyHG I$}(./k(_xQ~QJW/+q^Eɫt1\=`7l{Z.Iz<ۯF-ڥèxm(H32]ԣ | rVoTWjɋaB /](_->za{I޶ِ7/yZtj{ݵ$׆es 5o7Apq:rD-\=|26|ݷf 7~倸qp5OVB`Ogf#q} ;0i $ڵVYe T IFCGBk飫g-׺NC$:磖e$dt 2hBFzVcJpRxjkFf6#UB|vvH:sXh}Q<^P츌lTĈ١J9kԘM2'ZDc@D$Ǎ*:ȹ[oYOYnY{uxs. E aOdR TR``uIhAJYf*QTHj'V&C[(?f .Q? ݒp5sdާ$w樴 3RYzm㠬Lg[x/~77g]Xuzt6j+W.]**NErW>_inK:>+lzO8OutǟpDǟp**sYd9cH>jH5sE%EEK*XpN$-׌A6e y>rMHʠCS=)KQ1팜51~hvvqksyCXjҾP"yl ]@v4ALWīfGJKλ[N} Wdx;Xi$rJ)bV@OG}^pރ{&84Ds.!9I;EF mxa!礵|01 F2JЧf6TVGMLy;j{ؕ3lF:$̠Fh2m@NO$~ȝAV ۭm{t á#ɥCG̽+&_$ڷûj @+ M.ϻIMJ|E,Vh@UZA rЃ$uRRaACM6\Jͥ-{e.9?i~7ei4̨]?Sr06z4wM}37?ލG-1*U;.ZjFLXjm݀S[R=R8c:>OK8Uͥ ,xrPNSAN0W\)R+ b)c珩ZN9,26hm&LCǨb"Ι2Zk{@qjnI[#rڸ,HoQ;|ꁰ=yB'4' GO~Tq\ =>SSݻBѴ-5iMf*]xF]Xc0 sb2KQ0{^G굫 Q{9Qsȁ%QE441 BLޕrݺH.$Y"TJA,pZЏ00I댜]Ah6VzlqVe}l?K~|߻>-&5'҈a谦I󧚼ע xiQD%Xt"xPW Qeͅe8(zjgJ<{Bc(#SVPkMt+J9NIɥ! ˬS^18O*Ȍ\zkB]쌜34tV56vxA" "ioatFa|yzü̙V<3 dftr2_dt'fVۉ3lЙg*0H%/}ؠATmgB~fۻ9EĶͺeI׍XlH[$moyw3VzlzGNj<9욱[tXs˦^0cZ-YkD@`skR,Eq;* BUұ*yIZc_V9UơǓw/Zрْ QfO"ĢAK9%' Rr۲L 0,Jeŕsi ;묗IZ|4sDUǦԵZM?'PіRЩۻۇu˭uۓiiYuK`E!tx':bMF 8*[d);db|*1m{:10d)nX Zù!% 0JqWk*Y(5n4%ºSE&e.yIR1KyR#h„F4^HtOӫQ]3xĬut*FubQh*e$GJ@TŗM Ry(i /'* G{Jh eJ:z|yD(yƻh_.? ߤRϋkH,'Oh N>շ'-6ruR~S74`a˩ϔ !v]mOc+|=`l|νҌ3aڠ@tto9  CI\)\BFkU>=b]>vbw2М2 FΙy1ڤ0xyg2#xS[\[UuQT(dV:_nܥQ:X:hZT&|rɣ #i}W>̐=/ZNwuV'1ʇ{#]|W"P:/?:q(G9?Q_h?GۛkM17ȓm s8㣷q5|5KΑtˇ羨Ժ'~nl N5k!CˊL\Ƨ3v=jhzKHVAE'׵U/5L>;=O?L贎;zt:_})K/;/١_ן~lO?8NH7 9Jx6ƼGj|?=9kg悿󌼥9RGD~QѬ(-a!/}"$ p_MK{Mw7 5G5z<3!޴ ɥY f/U@x72ݙi=G-!_9GC0Ul\th5TB"S Z66#=¸te梄q?pO|XÞt&eշtbϠeXޟ?8>eggunԹsv%lGP&h WvϿ=JhWw(ࠋBg' S 3bf25OZPUr9[f&*eE8!4 tDzX/k^|xv0XPq@[k!' B6x%58kt9QB.T=&QSLz{!M9T^jljcpw՛|;g7KMLp,"tɎ&xo: *v1DT&V`V{J? Ym X-=$ơ%]ŘHPY[ Eˁl~HKVhcҥ2iKvf-GS'2އO\UKTEp>}nQ&0oOrޖMZ7=cyxX 3yUD`ZʮJ]mixsդtj0W\ +jl'qkU\5iƛ&%\Bsi̕lsu~[̕HM]\5)@_"㍿]-9ܖ|t2]r8B7_t :=2{ʌM}t~:ߚe4,>C֝)@DVyV.?~;-o ֘ini-l<)LB3m[dD`YG(H R\Fꃹz=9"s%#mN%5J\`^ZkZOg1`8:: 瓦rq&w-#W׳,gL?OmݥHjGoWoNy~>==uo1\$Bݫoٱ|=C0 r%A 8ܪ&K`˜U`o[v_(W[ Pg=R'.nvSЛ܅}gI98CmH~ s#`?HWn$VaV5mU\\1L ZdYG7: DE;c.ؠȧ-1}V6)BUstvo35⒮^bN7qzqu0 m5ˋ VAK !Ȥ T ܨ8}AWyqn@9BQ9t}pz&9*³:a#Ю:b,b7Y4f{$ Y?d؜SMVA1CJՠʁbJ֩$ȉՀ"mRp'+Dž_w{H%ؽiVkӦ܅ s[)O˲=O?0S6(XSruZYER1Ěj1-e:#|'YJT5 Mg8VK}ymkP^10gbd )@)*瘱]lF>X_121dɗt(A)WgMFYzacqf Q&Gr 2 'VStlK:A(z Q%M JkvyjϭO\c{ g'Pu3'KRX LBo ǢɷT:TEGW S[}1dy"|xVn fHRdL X"v-p+"r-}gPTؐvOh fy"NϧTқlZP )t_6{&ǜk{wl\7/P{ywlRgtOQc4$aȱV|Cl#vІvvvFvPSقVj49@MuX SPq*F_iG]t̚' ۪#36`LJznd/bv?¡.NizօƣvOK"_*Y1^}&۪,M`=9 ۓE6=A<2kFG>bjewRGvZ{ *}HH]Hһ⌆tIJP;Ў DHP(6Tb}"!f}Y !GN=˜=+.J]M,A^Ȝ䄩zv@>,{zQ߳OXLkj ez~&wjx:]I}:]w.{2'Qx;!HnwbȢEĄٵ$v thTXkO hIֻg j3MOGчlZTHV7뷂/<Z Fuh]HJě .ww?g|g=i3y鰌GS-2 EV?-:r OHvbduzE6iS{Zor)޸Cez1x&C E@"R| Y&`Hx\yy\{Xmzc>tO#)CpeB,NdBZ} TT-9tPģVcSJUYgP聵gd"@成(d!Aϓ7qėDc@u4z`&K];n#kqMvǮ9:v1CV5uO ܋ +Zh+ew_xk;1*h-v^Ya Bg' Sx+uȔQI1ՊR,QLD2@TJZ,mp]/Nly#/!FGuX_+з^|nT!`‶BNVvlh Lp4r\"{LV ɻ@o/$iћWd [ث6w;ؕ]oX}`iӞIe\yc:#mޛ] :UwIEDdm}UiJ2i7TDO=~5/23l@뼽 _4Pfӕa f/vfU)շ\3yx{=/hCwޅw=[:6vYyuW[n8)m>M7m&6{jtHe?ӶӦir޹$ark^#<;X4 J`z2F+\R{wPڋviWkAzGvԄZKnUԌF <C-UԎ<sKéx'Wak\yfDI尥q)[rVRgH8q?t'nT[.d5[Qo8[AF$U%v:b{sʰ[#^%Iow<U=ۏ@cힱƽO}bl\k/&L_?s_K|7H=%8obON0wo3CVrC#D=cc]^h?8WeK)a@r ٵbLQ[;?+gaY(FSD85ia9HTa=9'üU]*qQN72> }(=T}66%-emB^_>pD)KIm`Gљ"4Iu larP{&/[gUZ_KՓoofӛe11Lω58_TJb^#p4~6-/G^'v_0Fu$׏t6 iZf > G%>z47],^c:*AGm^mweJ-F->4>_&.u=+B6Ys|QeӣN5W:?V7(oJ}s}{櫷W]\ 8nIawOF-͜q9^f>/K/Ft'KVꧭ8 g  p[]{X?-aiy4!\|p؀km#[løXt-~p\)؂ݭ/!DtQq]7(^"(HJ6G,[DImH14bꩴ}t!0,w;" &XӘT8'S;pK) p^yC",UL@ PjtRp X)c<"ct<)t6~&&uy3X3 ӏT:L~sNv>Ṕ:^:QNa;/#Շ g|!ce\g@uQKuT̙"7<) L(~7777E(BiکPڋEǩFxF@KaApt4zpɯf0}gRs"n8JOX2DK :y#t`2T`gu.َv4K&池vgܱ/j.Q{K'h9d&Re)'!Gt0 `Rʹ2UZz)F.9f`!f Xr1XLd#FutKКX3ILjHNxB٢}4FsS =جzk!DT@bܚ$#SD #$$"pFBb Òpz #4(&LR#1s&ZLj9; ph3.EǸHOxōQ]L[j2xsFI){Vyk4Y}sQ'\<.M_Wܱ/ʎpϳ@XY_1glS)䂾x'\[zyO#W-4j5{yOR#+P0O Z(})ռk[f]h[nE(ǒ؜qsι:cY86GW:O<{#?,-~}iI:J)R %Ұha"*OYRRE4̀1o=#X`T9F$㑅H>H=׌@ FрG!eLDc'Gg>少~/7cgz#mF i^}?9xwjWڦ{' 1EQjBb7j*%ܒ("zB B嚍Opf;D"b^qԘjޛxJAP$/;,A'{&CD{fhH @ U&"Ŝ(%]:#g;0 'Xr)iI n-,j(RûMT Rb 9mBRvO0FB$89u"B2DEN &V9!<<(sa#(`%mhqw{%hJEwBN' _/ hp0rz'qaB( ŒpvܤD c6.#{Τ(+ЗºZGCR$kˮoؔQݝȎ.@< pq;`2Eh$hT1`J#gɟ" @ Ɵ{AjV"iΑDoy=U%/K2Q6w3Jc[E`4ʓiFoFo~3ǿXi̝8gӹHH *"eEM !As~J˨̕ddWYg_|^3{.m%yOA-0ďi˾MX?6E"<GmkųŤ N uQ+{tڿmDI\IꞦetDyIR(ƌT4qEQ_<&|]EM.mh71l:z}rJaSq|6?(qd=&5$ԦBT]jwh6,po0xҿ[Gx$շZ W-T[nx;6~7{L{CQQ'{`7 ԘeOn0KDDgg_N=smyV׶Zzʃ s A0 l~+tR;ĬW逰  |J~^sXeyW㛩\֝hMXDO/1V)`,C1c<U uOmFS6vuj{3wu׎~l5=;&t-e=L7kSp1z2݋9x/u'7#/٤#,x;nt @i.$Lgz|(}#Q~ʛ|ɗJlNj91\AFK"($Sb)H/82\K+FӡAٔ]Ix{KWn ؅z9'?=_<[޶y;`ReWSz?|ij3 5nR>qMN߽|u1+Mc&*C צ4QFyps4trJ-[bN-mЇNW2:#hnx)]1జ"FZRf( ]!]%㼶 +엣nX.>F2);t9⳹^8zez_:t4t`*=s^[mZ]bn1thvrb ]!]K:CW ڹt(:B6$DW v=]YRK+F7󾜮 BWGHWD BNW 7:GIW[:*>1/BU)Oi7DӎJRhFw4(ʣiԸypb10t<= gC~vT9OgwqD9|󼷼wC(Sk _|9$OݻuT{42/.!& n0;Wwzw zs=8I6Ƹ}/pOKɔ~s{m>?ydn{m*P/ʼ#/dg=Gy"Cy|-04󧞹Eymfs~w>G廯~I=#lQfѵe)ڗDުu^WY%(͍bN">.>C{yz沷3v?|(r}z7p){3J%=|K()2*T|q:+gyy(%7\JHB:vZ*A u֪l fjoM\';:[;cH՗; "ZJQ+ω6{Gb:0JHkf8WHnLCQ{X1z, mM!gT]lJvl 䁈:u))Ģ-&7ft aہaryh0%/@GmlqVH blSN,UhE ռB:kϢ;;kFFP[Wy'Qk4GoMu^I"`!-!ZvH¦lpݢ> f՜%C8˾L Woq WR1y\ϕIv| O`0u gllfѳq5eoPYf1h\ ?k1i3jFq5R 2j31frrӣ[*3&5y &% xKd䰮T9 ]Qn$}qPD;D;Lw[zP낂0dj*g':RC֡7q1옍}u`n?y xMRP'w d!G!ϺCE7 F-)~Ba E5FHKCH4pY/xp`\Sclci0Ih j hN\76xk+fnQ4 C5kփ*H]6 |t~f҃d &S@rZx tmKzgAц?jo9hE kJ (Bu(AdJ 7bdzldNҳ֞fJOQ!,yJi]0d@b~pQi4 65fs˥ZpUY.!b!c-ٱP|Z/") %;)Z͘-, \g'~b!(E%蕲L0H5v[x#wn/nnn ,.uFW u1AZiTw[>8ݻWΰuJDcBXfq_~7~xo>bVRl7ح.jp rmҮwt/.~Xm5b}7c7os'?_\]㝾~u6_} {o+p{Ϯ~y:>OV?;Nٺ11EhzS>;E7m=^m0~i $&ʨ_t81__ઈ3`INhZq3/ S@GJH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 tN Kr0i'5f)N FK$ܻN@) $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@:^'2`':'pN k:F'FEqH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 D$'Drq8֚w%8 8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@z5.yw[M՛}snCvv^|8~ K\gnZq h-qmb\: gC#^]U~1tpbn\c6:]1Jb"8b>.mLNW@icN`buel3tީ7*%k&t4W'$wPtjBWQ.pL+~CWBWHWVwb &8k;ɤ%LXci-QCsCR]`#X2tpH&m+DiXGWHWVT{pv4{o!|*t/cv'%**$ulL|15l1Jkl1N*T.|22A^_ʖu!32BC>T0 vmvcͲ࠯gY"] q} J߱O :{dwH}oq8¯UҺi"VDs ZRY"'!ZCdIF6'XYQYȄ'6en<+X9ʒPYen,K-H91Weօ"m\<BdY*:˞dC$ƜEIb3QeBrDqmJP+DJ#؊d\l+)IFFJw.ؒ2HBtKg3"hl;]_ttut%p+XRGxpU3ܐ ceYlYBK<="zlX& 5G`324G:ǧ.RYI4VMEU.45g鋤iM)9Zq.+DlGWHWQ]!`]!\hmb%# +˸:!R3fp_uIthm=]!J9]MOx!o>`Czp-mjU)z(+z[AW]ю5=eZ]QNJ֩_vBututg_1]`ED2tpL-m+D)uGWHW%MtUɭt(U ^"] nuRcW0 ]!\3h_u=]] ]Ilqx,Q6r@H_* SnL4F,)D5QNU^"M+c"%B$CW6k\=S .wj+\]v*B@bD2fD L1 \ ]!چR.a$+D~u(E ~;t4=ii~`>\0]CPxjXSEpT-d J ]!ZNWm] ]1%IglY2tp)aet(+ˡ+n9]!\*R+Di QV6h ֭>ܧVa@;D*zveJrqU!4,[RlBHղY9jYQn RWXtpM2a1ŵt(YF;JWd dH Z+Dio}bҬqXBW5<^măZ(Mh5wtue  ]\NH*thZUe6{U։)+e: d5]] ]Ȅ K+DWZ("]] ]IS+QW$CWvBTvtut%80>`O0xbski,$Of454(۶k\GӯBӚh]`Y2tpJ-o*J9HWRE&up5IF]!ZD)tGWHWr%RRWX'DWZW*thEoā}B5kA7aP=zhuCk%4-75Jttu)V t d2NWm] ]1ƴ nT bNWꎮ.8¤ `ҡ+U*thmA@ %MI[~pFrUbsxLhŒ9xFs~Y0KUs~#ԪV.TJmWMS9eWJZV;ޡf& U,vPB:]l`?fSoRd d&0h|+䦣K+CmZZ1ATq7E ]+P/'EXh M\Q'4UgFgD\[)+ 4BBWw9ꂎ/ Kj~ r.B2'Ѷ4 ++tBt)ID+Dm QJ7CW@>phZeՂHCRWPʖ-5Jvtu驤Ę S-+µв#RҎ.Tل 7G=BW֨䤣K+ )+,h:tpIm+@))JPƪ-r)4`D2zW&wRpEb6eӑL RBNKmW =ͻnׯz߽{\8@a(u4+JeeVqountǿ˜s? 4_m22 q]˷|cЍŰ+Iᮛo؛ pc,VlGl`hvܼfHx/n^w&|is5ž4{XkWַͽφdpuerg3ÑTեBZ!:}vKGoc;Vsl+}7wWk ^-7j1_ath[kz"v6? ]>ڢ'޹|y!Wہea>7f I&Z}"ZBet%N1sY!VHEXAڈW>cA vYf)͙ęAɩ΁JbNPٙ<߈7kt;AT]3%5flM jg;$LZЦ I7e|=/~eZ֧/Sy2UgVyO[ FlKXQ( = ?㸷t=m? 9x]_~\5Ev!O/ |ZDc/TὃSU #X/,܏;^8jP1RntKNJ;IՔ,d"\NX*kKqZDɻAub1'drF~.Ί3x"D\JYyntQP) *ɭ5s Ot wָoA=Ozߏ]( v8VuJ;ϴ!hR/x]>RşQQXC&-"ɔ,0&u㺨.jQiTmbV(|ቖRp˴Y\;iA;ssP.%eKu1Bm!(3iN1,5ô҄^v68χCg[ Ă`Ǟ0y+g1?Wjg->:ٌĔ}xIe?;N `$*Vez-+e/$/uj۾%Bgߒ_^"p(bt932AW  w7^*P\gZrLq^y@7jc <幥BD 2X:;+s:XY^ ΁C{4c[x'DI'=Dp2=`+r.xtG=>Uz:!֏p'Ӈ_oKuW@ky2Br^NgW_wP8y7#{_aTJU࢜ë3(+`Ʉu: N2Xĭ9ލz3y˱ キ f<>ޣ'HPssrƴ|y:0vm)ރP8uL:-s!`*'ܗ%r$5E ρJW9?&~>U 9WRdO~g|HB~o5Ud*5#HJ=1zFnLDA،m5qlc6``RJS-{hI"tPx F΃L(Em1+|kZb\V73k]lQjz>p%<7 b.0 oys=K#p4jD -&vpBu_{X [ۍ (Q`C|pi>ѓ> *^\ڱr{8ܱ@~4WXR!j>C*'~m+߿..sb௱v8"u_A.GAxĴܙN~\KuͻYiKz[Ke!*ɯ_B4; Jx( l'1NZo#_X) [N(r§OxvwAZƅFz͆9ԍ|> zw4ܦ';鰄iФK-)([2=a,cug:;L6KCč}.83wabέAgjUUxjPDVZ{Z-*$Q!H$!GHQ`ȴ,#ELKk.ʺ׍/z}OX'fy.rb,)Epu1* i5A*fIN/<7Q$n4Ȧt5L{BFl@`W`pEZ<+ξy&S/{&༠MCq^ JXPsK\c5IIO7UUcz>Us'j6U5Qh4t=N# TI}$L3)P4VHOuXHb(t0҂!Ɓ-|7Npϟb_zsrTVo"ة>ϥbKD/?_n,Ba6c]xҴaHqEDO@@_ߜg5Mۦ _p}Gc*B{z5?L\{q D|1V",$RuN@TRjgwن7i_:ฺ]mGa-1_XŶ0IÔBJ " vy3K9U`VC/-+-DId6aMv~&{6P 6('L2Nڲ@ \j*s/.(1i'Zʟ$/J+g(b}0)#T>$Hx+Dż^ `Q9!IW#|sV5}Hg'R|a>ʗ1,9hƔ X4)O < cG ^(%J.i0. iƞ s XMԎN HzHDP*Wۍa<.6-ڻvhxaƀRl1 51J"ڭB8'{gI;mTE0 ,>XP2j!*Ě,2%*1#օ$hTGema<,Fx cWDdu!&)pPT A< 8+ZPˤPͅYFlQD4"2 $(`,eqaQFID=ZҌ+rϜ/9[DG|qq2/vH{ŸdW\q:\pqc+nXA$J>@"tpN{O#34kA(pqx0JqǮx g?{h_1Uq\L 4,*Ҽˏ熗pzt$HMs5 jKE@%yx"Jۣ.w u䋚N$e)$!>9,Oa:&YKO KhͽB&-jǮ2A). vh'T p&eF",Z @'DLa9[:yxS(RҰ:1͹$v_:Ǹ~H5KG0' |A"@bB}&Ҕd|L ΀cIzЬglC R+%ƒ*1e(8 &JpQ*JQ K)Z$!jqû95y]yk|5؁-6?ft Exu4IR_,( a(fFV?_enlѤzs gw6zxzxtuY2л:͟&={Wayr~"%~u{2xZ\ޟ?^)6a'6*U{h͑k A+YO7sW߯?c[/q3|CFuvu7Tz\0\A(W852} R.~N^ո=.'Қ?V3ruRG"* O_sa|(e+IiZƓ ܑWIؒN<QE (ۋ 7c&d^ 37~9[q\]v|1T7(r5GjH9գ*F&լs]ɻ:k}vrht7) 4nlNawcy _9$ݓGQc_v U>QTd3<։JfqHM\F2\}[1^@ҡDi]3[V//Wm$ä 9)y`Ap28%x5pv>pF̽b&'&3aɄF/EMFpUݨ۲5eג;j ^/=#By;sXi3iӇ4\H S)r'Z[ޙ؍;kJ2Vzm x iFd !rTEpεxG ([HAh&6F Nsg@D"F!:i P@az%LIMwY~`CE2;ӘL| g[hNHlmI%)޺ν/2.,֧^t#6KCqU&3%Hp]Xh R*#XS4"(΅f"X]̓[CNX B8*b!$iNzUۻg_),/nEpgqiD5:(1Z|д#2N5 9)4`&8qsr˜Ar)ٹ.ddf"si$QQtD@$OX0#>z'4 -).Я1ԯ.G{ i E>>BJ-pQ͉FV yN!pZƖ[q|Z8hA;zv>:Qňق;ǫ0?j.AY(gޢ\`xB7gY< ,o ޻(@6zAA(E(CBB"RR8za9zW?) \]osZ^#VqxɅb< 9C<1$BG#ArAV+NTk#9NcF1Zl,Asglr",%! jCen Ug}~<͖q\gs<]k a>-P(lBgEG_^K2d@15nxSp#եkT'[z,5\ZJBuFُWEZ8+ Ug˼6lV ,W$< YIիrQ,a%2hl5^g^J1]%\t\IBW RNWRbpOW/dHTv\BW NW +-¢Ct Jp ]Z]nrtP^]}=t%iz1@q^%''RWJAWCMvp4:CW .Q]?OW %G=]@"Jj*:DW$@;CW .]UB)UOW/Du0e \Jh8wJ(u^&]1r<)@.̞-z$WS?#הOý=q~lOg.WKq.c$^U/\k*& =a*yג)E 0Ï12GDFnpeme<`s-ΉuXQVƈV>ʏښimt9e#oԛIyCԖV0#)6gF\󴟢1A@F bYQgסmz5&KDD1Fs}Ta 2Frό#M%E\t"Zu(1yI $u%1hH(/01\1[iUS΋bYYCPKA2G Oey9TQ#hw^o69`ѝm+;C fvR˞_ MKu`-;CWWJhUB)tOW/nna~tRJh0OBȃtrJ#YALAEW .`J(K'tu4 > ps^p5:qj?DCl"܃dOWcE cN>Tz?\uZ)ϝJ%{ztEšw09o+t>t=%mE% X s՝`0՝ gJ/at\»BW -?{uP +N}U|%*{%*B+=wJ(qztŘl 2Ejp|Gʽ`ڊuyv=o+f˭XVAQcD39\ߗl6M&a>$xB3 ۛa67-Y c ]@r-oK3Ù@B#ʴcǥAIyd bpDsĭ!YqT{0 EVU3fJJxuq =Ik%fb݃WW>s w6b_|ޏͰ@LՒn7$q(K2ƾYEf.osVi9(ET~mWےϓ?@ÚO'f R{7{[f^7Bf2ڂ{%?7Kv _ pu@Fz[+z7ջeOWIfݴ?o6z7_-||u7vIJ1 &c(+4[c1$7pLd GK!>:M\cFD )n J+Q8 ܤT-́">\g*fC L{qi# +,z̝vX-"ro0CJf/hӨvRK2Fxc6h :(=aqN|7]>u5@=P?%d1==ٲ>o@l Ň٨ʫ$_bw9[,\Rҁѻn~.֯3n\s $Bo kc:g0az~jN 4(*Nxk2B @$ 0U^մ Te_]B[eP\\sv^?~ ecR(7{=xpj&e(y𣬒6L&)@l!+j3X/X [˛Adh[uJY1ހ!j< d4EJl|",]84ꍲQPv+MPjA˛̮5E {(wa<{k|kz$l/e[B!fA;($` 8i:}B*Hd8IfHD:1,Rg2DO[1E4!)Hno$xGI4 GZv<ĵꚬJ՝3ct3'hh8~?Ǎ1^A E]h;͘]bW X{`V70S+irǽ>O}c ԧJy=ߓFqOw)lvK% d&QQacтϽNQi!1CґGbnihU"a!U $P>sm Fb>RƁJI頥pTcb:DO7S׷HQ{ymya_zPzЫ8 {GjבvS8$* Qb'9V)Zk3FK4jpkÍ0may»<*=c"^WƅoF_2s-i$eFL0;b4364S*&BԈb;̮_cؿ;hRg!` HID0RL,ᣝ ޵q$X~|,[W)̗9l9~3|J%OH2]SzLoBOgJ`pVtb + K r []/A]8_߇1B+ p HknUe#JYzvh3-(hR8|9qG_,Ũ:Ngzw *{#!px[8q+0 .6Iew.F˅Ԝ[qw~~y;wK}nΆQrl:2$8 uϿ)Aݬ#,wdzw}#ir ͇m9rz;E)7q`Kw}_A"\HlP]D}ti{`Dɪ~Q?۬#úo\GC#J+MGgr, @4&k70i#=a.E{2aٞw:$ QEA Q1^J'G΀!QKk2#B0V{N&#ٷ16vla6,)B%TE@1Ǎk]@D0VZ zJqTQܦ:2,"1rbZ PKcn=.9(tL-~ ]6N(3G[Tyĵ%ՋK1v_}ԲLЄ˃>p^AUNGE- [Ƣc\YL"` Pʅfu '1beP4:62M-uJ/!ɬ IUHϼR1qV{1gz4‚SѷIfd^Agg$8K&YDLϋǒvlVNV{&QR좞uKe?5@=)]`}H9$Vkk5@ ya>x1"9HiDh j>Xkt>D a;e:KMbPO%s>>)Lr0·J;A:$}̤v#! kN-kǷ}$E✊-Gfc [+ l@&g4妓a'@}Mwtt&:0od|Seh11fG7]A*sX1+.ʛr WhdɆ ڱ߸g;r[ .F&5N~T]Ť5@ZWOa9Nm$ͫYu- hY̸un7-w>ޡ畖a<notrλ<nk^uKAw;np+ސNgϛ!iV})mjԻR-Ml!V0mMF$b']qk IQA)ER m|,R;嫷HoxjK÷^{z5][)>zV)~m%gD?{,cFj)BybFeNxbbU˾H`_3Uz}uxiʂYqƔ&o=3IόE#̘U> όhF!$R1=\z| )N Mt 3r bG)~/\*<&}|CL.!JeQ I甊{c $JADIJq8>ɺQCRT9 Q7/ qfaHB9UaaR"2ollߨr\e-:o;$CJl|gJy9jZw~&nLcetgNu_c8\8NG_Of'4}@OVZ- '%.t$҂ kܕ_St{C_zr3t^3^x.m| <-郧-0oIw~:VΆp4<[8N}.x6U;)?MxͶ?9w&WIKX'W<A?7cvնAʏ2e+*,Uc?՜3 dةR;dMN۾ 2KlXD0SV-g[g.t`4^]w$B S٨ApdVl+=sciy杼&%;êZMVn~nSRFV̝!rhPVD@(#8HCBp@nY1yp昽u2]Yev3z!mv61) F8!FcD"2,AqjA/ %t񲨬67~+$m^ٻ6$Wc=Rwy%/ѯ/b_gHj$Z_lj^z˕JyoUOJphf@a1uʔEIb%3QuʴīEdwNN(fD4mmS*,g/OVnq_Ⱦ˻+o5nһv.8q eNFD XL1ㆃNmr1H 5 =^$=W=נOĬ `OzK9BI!l"ǫNۆ2% Η((Iަ:eA"f ۻ])ߟ]7²:Jfr3Zn4n<:[6P-0 :BD8&#w=+(4LjMm1q㻽pX8Q$h@=MHhgpZF 6ݣ1l>3@HB&|5qE`1?WJnWYƳGZ.L z 8Z^8Gt.:0f n+eHg 5QBcHA3"$!AȋW:d9385J/cn,Q93 FDZR6$a,$@LR eeƂqhNX+0ww b̧suw)gRGJLRJFuL&hC ݈JFTz"ˍt1SLWDZGG)qe O@cJK#/ra>U!b_%'Gصn~PګiQa U`[, 1YNk4 bAkF2`vgAETXj#G,Brd*Pϣ*bq>H#BUh_*tQJqHW\{DWU /t%#tォ@W/@hm=*{"U"aG5z#x54Pqڔڪ_L4ah?^&Sh?1NODck 46D!cF3D(̂xb}l:efԪ;4@TRx$s^0Fʒ2~rRni rDJι+U ̅Um !#Ixr2{R %𢜎/Vˤ9hn5v AKC0.1LXyc ܪ\m znEBJdBl,M*KтQJvClnfo-ng=T.>oߗou q<|+ "O$?!ɟ~b<7, ZYPr"6z$ ܽyҧ ؈д@Hoh:zi:l~94ӧ(*3\ٛVNW%>ܼHPf.<s ɫ }j:|Qew;Sk,}& +OłJ !2ik~vݵ:Ƈa=q<ɱNNUO%q62WkYRrWA{UCME`^ƋL 7W0U> ww鏊%-ףQ0N˼x'Y#ª^Q"/‰͟h0 T03?GXNQ6[ {rף/9 mwvZ;SVJ8Wv*N{q弧c_(yR9)']m5ozdZGss~x6fSfKQ޾]Gn?|K|I+׽ig?5X {< ֍W ~Q^_/KV[Y]3aubC Ym+qGj;qLV3hP`oGWg_ P+6ѵ^'eupk3}>vlق~!۫l/w,Wf@COۼhC=& (KI䅋Y]ȹK$UCD*Mx >1N\Ň\x?&}i^^S2}Ɠ]BH8OYsGa2n܇;{g E5#vR61ge:jG7x:mXp ◚x< ipK)]du4JQ^w._;Z4| sr'A s\3Gi%y4hxmѕ炑8a)&?]t^:VɯY+暁Z ĵ&6r-qU8.Yy +yDG,)(KYYk+WMnʛwrVfFŭ<*;U p%Q}IhRR^b QҨ>%#` WUFFF偮^ ]MU0)dUF {_ QCegDWRpD4F=9]= QO]&qp|ZzZx`Q(YEx]6=H#Bio*/tjw(4]!`Uo*Õ/tt(E큮^]qä=+wEWnADw(< +a"}2]e&r*d@W/^V%%{ ').Ʈv1Osv Mg7svM録1|C ߃do*5/th{%;K+ Z>p2\՛9vfE%ҕQcTsx1P\$%\ڎ`޽st}Q޴Y-Ǫt uҩ$IIu(,-IP@ w:,,nEcjdSycٖT> DX қ(}Z2J~}溱 Lj}3 ,J@W Ҍzy6;y@T@b^jˣddTHG@,Kܧ2)O۬t.0~VC6jUW_t|h[vzSmVok;QlٓۨݍL]Qubu\U3x*q*y eCeZ-%19j([wv{VĴM7~c&EK*~?߇NN=_O)tT2A⫈)|Eケ|o 6`o$7"(~4r!˫`5QggUĊγ?]8:ߝs8/TWѼ3Ss'.W5UנWjpN:gFΝśF˫ Bxy۳r2q^\ I0ϫ-.L_a z~8ś g1c Q_L0ܨ#9Hw7srϴ8Edp5FUKϜ%⎆*qqZQ.rʇ;vĴAT$ʹoܠowNst!!&Gm>fs&ɍO0FvGGDON-L˄I[Ι4)Rex J#z%2*O-x\ /t\PəA[9>pj.%e]k4.!~E'%K=!@ IN#n+9 p%rZ"4H@w5ĵMpCI)v!sbIs3 I'(;jVu6c#%%;{3ՠZ$|BD*9 AkT# 2Q)bf+CJWPŒP]1QC'P1 b`%D @'Y8S+bc9oWpu.*]JIJJ;N)"$#BС'% #6tmU[:AI4fcbLQf1 o0`) O䝵DU>5^bXd4*4.PmWH2is}N{Gmb0˨s E%ryKg4y q>I!% ۬'(~FLLY=!4(B$b5NF Ib><>n21XO[jCg2 *r_|F=o@UyItI(1i))8xKJ$$Pghu/%z2aZP(EݱF. 5mu1|nQ  -`h)%XGV!E#H$jFK>odž8Q$hʔ $`ўqs$hk>$"7&]n$vkgw0ʂEbU4حl&ɮN =kP s1@"KN\Qs(J[2JB ")cq^$ElOc+ y :ΒwTB\1`!MM((py2p⾧HVGq0(C/^e)EL\:p,,"L#w/g !./5GȳBmTzr%Ōq2M!̓V;@CX !`gEhH}UTUR: xDw.U[c tX,`X&s\"1(l1IeSI1δr*ٴ2@}+XEer L#+%po]MrF%JE5nXU]WV1(l rABj#2Z&SW{ $!) 5ID^,.zEBb}UTXQx %EOc`"@S[2,pZxZ댉p0 d3i8w.h*F%̷<k1 E%eA; B u@0- |a )DTswJoi]:lSGB$cK/(yi:0J[J'.  -*=``j>pLqΠ&icR] cA1J#H*H2+BiO,Q<{ ȋ ":t@2GM8ڂ2vESU_~(Ӭ(al2JRpic0vnfsR ("m'PG]rs@ QMހwK }Y] P9B(5b2H=]hv)0ְc[8 9@g@!t9Nhs^wŌ:hܱXK4iQip|nhImGhK;m,tM39(k$`?Aj؁0"ep% ~`uaE^Y` E*C DYl9i M;mb9K4YF%Gj3Ko^ujJ*[fy/z+뤱^u039l@E}׌&J X{0sɺ6Zĩwwv ,<v9[.Ln @][,qtU  G6T0 0o4,.,hZ\@Gɪas7㺃)+M0 |ျ!(A/Q7C|̨C Fr{Pڄۖ?샩=xYk$}(8d2 X:da 5!Wў iDF9|7'ޕ{ ̠!GX6 8N*t+ /@$\0 槍9ml!&\Ko&".ńE@ݰ DS(]p9D LBF t^\רtEC@Tv"#DoaL`j?nZZ#Obn\Z@vJԳ}ӏzUX}RJWlqnѶ͝{3?fo)N~!źd%0%>%-N֗n6+}|5[}y!Fv9NO~^,W8ju<} ~߷s{k+\\e;|yxOrV2?=hlKrORV' 8^|u20~'TZvA_=T߾cPql%<ԕrv;]Ymw-6":ŰͻcYӽ^rj{G,ȍi@TOfAZ7'\IJkN_c;6_.Zmyleq}dyi3o*E#,>-(V|ꟛ ' |0OaD* >؁ CI1@|^d**Hrd1韯go>9)}LדּWFo|x+8:\G~ʶZίWs_սy7/>t탟lw}͒YP\/wWym%_yvyr'ZŞ8}jw7'+@r 2݃Jcl(Qbሎ̥O @WDl鱃1@2y`oj.Wtm*g=d+Wr`VG),>^͕CdM\|Ԏ}kЎ>vw: [s3sd`GzϛtYfwEY4/:x}}C:25=8\ƐZʟ(E "ȴUCT) `yŹhns۱VÄ0aG_Ä#myJB&&g5 Ӧ>9knAB`Qhm&{t:tdyύ=R}nj}QYyUT8hAA/~;kt .Ig" M25hEj&ꔲ"60yԕK/_]n>vp1} =_P7=$]E%]TE%]TE%]TE%]TE%]TE%]TE%]TE%]TE%]TE%]TE%]TE%]TE%]TE%]TE%]TE%]TE%]TE%]TE%]TE%]TE%]TE%]TE%]W~k_Vٞн h^TޢO&pß} F&|-P)6(q#[kZih$<@N?p4&}ڼ4V6g>Aɹ|{8k$HQE_KXKX&EF!=^Dƈ/{=y ];ǔp:=5~eIHuR )m{ޅـKo?kw_DХSï7|m.?XWd \f_\g-d7**kڈlmN Wu>X[h̹?oY?@_.:㫽` B;J`^Aϴ֞r{S,13Paبm(Zfհ8(R)ơ"%YjHΉoi/o!c&J˓ϳ|,w8=/yXP~ڞ1{1[q%zLM܊Pה65*.mU;~nrVN{!yL豜3\dZ}˨ioc<߽Yn:0_|Od5r[^b+)GִͩyzWgy_k;:g5n)#zP)%HL]HE]$E]$E]$E]$E]$E]$E]$E]$E]$E]$E]$E]$E]$E]$E]$E]$E]$E]$E]$E]$E]$E]$E]$d 7{M]#.+jvJeH-v\pu82X3Ri; V6fP!F/ɷ6DՏ!O9]%u=pבr18|MR'!Ym͛,2V2_Ah \+g2$C w(ᮋQ 5rCc\uX;5|Bjiw%&s^=1eE_"NQꕖE,j_29um|oP fΎTBt`Jq|p=r|O_t/_uaY1dq"-bL*V9nHłT,HłT,HłT,HłT,HłT,HłT,HłT,HłT,HłT,HłT,HłT,HłT,HłT,HłT,HłT,HłT,HłT,HłT,HłT,HłT,HłT,HłT,HłT,ޮpa%DžRVCo4.8b`ઠ%cLS_o#WQ7 ̞0 \_pf4@0Fh^VQ(q jo4Tup88͜G7y EW/Ssb}}t*ϋ|SψWҞfD{FtW%&gDA >^ּ&5Q]ܾ5^TTޢꚇz[CMo~0hǞX[ڶ+sRLt06ƊzH } 7;=?9!+,ebM.oZ aZT@%hYK \f\hS>ծ̲GotGf͜0 k4)pضa#qV ਖ਼evc+}ő3x']!]!]!]!]!]!]!]!]!]!]!]!]!]!]!]!]!]!]!]!]!]!]y;>_{P{[_~VӃÃMC_S)y?GAS{$7ǽcF>X1?X1:\c޵·`_ȾȮ`"D2-jx)qHjJ=l˜fwuOuWuهZ^:{*Rd+uϨ){mֽ߫Çu(r8uҸ(t"XU1c8+< T/l˽w,=#EbWzX 3+ f{&kٓSJ(@hfwS}MX^c=:X3>9uJ'^[~װ/^ |;|·]zrvU, } qPiЦg0H/DBST !@*Q+ HyL٦='V5r}O.z7搚YV:o ϼQf9b`g?W֤%TuyYQ{:g*ExLZyP 5sOQLYםQ__,p"MEkpb#}5@}HA΄:z'a}fOͿї_\H۳<5а1dž6o~q 8l{x0odX=}N`:4u aӿO"B&>V{uKY͖m9+'|^ q-Ⱎy0DϬDYȨ V,X xes" PK '-Om+Eut g$S{4V묳U[_Ayaat}wpӠ #p z|ZBES)N T78i|ӛOcGA'ʊ*`*a|椒:i,w߇z5]YTGm`Ətޛ- ׫ѰwIl" ੺0>h {|nzM#NpYLY w~=u|ЇgdT4Dy?ʽ%vX-h=d*>➙W4r7a'+LIbt ht&҃:8G6 ^Xlw 92x}wp7-7~+UwsKX^K:{3Ix5^/ la[u3l]M[osXotfaWZVвqnZG=7̇{Znh7jsޅA og9|KӃ:9=Ҭ[Xs>y͝o+2r%mzpa:k[ klmI4-a:KX0#-mYN$<=ϦvU#U\ S+avCZ9n^*XDZybPfқlF/166Z2ϵO s.gu&jxJ穨D炆bl]Gt ;+,wCGJsw3mqx7~ PmeFN[^JEN79 le&l0QT:xL@&Ĝp;M b˞{܏wQ0O4 L]wLվf݅di$brKi5,&UJ'ު cWan:ni wz%iXĉiVDf]:`q)WJ# =8UTv#s3:Ƙ( fMU'17V*RLMQL3/@;^3Iۯ@<]DɗtBҤgakhڿ΀)[zV}jNQ)8۸S?>.0ۼL-"39DpJmr?o8p"Χ7UFAaUP(C%vrMN>a 7Étw'IjlQ$9ѣz!ɻ %JmzrczAnu뇏bpSk)ЗBuzz׮$MqggzI8A}vh~4-IN 3Ŵ p|{bBV3kݛ6G1Qs11]syt1]Wnu8Lhz!m#i:Gj#IlؘF>uz4՟luTF6:dۨs܏Z>4yﯣqoSzB{uujЩ״pQ'û?|Ӈ?0KM?w'54jN?,q~nү>{? &{Y:A8O\M^u~Y$hA]߈$$Z%Cs#M*90.U%2_gpb9^@ 9Ҏ ^)K!Tts*E [2vѾI (_$ErD+ϞYVjmLc%t9T\{B"x,h#I;lH+"sb:ei0R JHΔboS)P( 3#gV2qr:KMSa^I:`W4> au/'$p֧4*{%ͤ{k*bNr L.Uiz -DsF>Kl"}҂o[1 B6*Z驎9o|WLj8p`0>ZJK z|vH땐\'wqzʻAq,TWlo ,6'+Rr,*1K9OB!s/ˆdLdۋB9HSQ0R /mJMgxfA`$gkv9BDQ$,-" vec={N|!ҔW sbgn8mdT1'/+MЂFMh05U Ifh&KX! D3.ATyĄ97j4n ,M㇩ޏ!f% lI},ez7u+VT*7Q3ɖ0q阭gѥhr6;JpTiX풱V)& q^>SVNTP 2n>ܱC}Wɠn_i /~=%R f"Nz$''"K2($ŃIRDl Ee,7< XT\d{8C ((ᕈZN'1XCT:.hTcŸc_MKJEl[ 2Mv^`n8ͰQ@o\J/1xN؜7VGZTf%!2dDSI3ApLG8$Rsa1rFG`<X?DdD$&Wk܏R©JI yNH"lP#wG\Qh VG/TSe匏FN*)y2!&̈́NSQ#gNjڃ\YK\\ƌlz:EihmFJDMxI?RL'!&__)Wpϻ5wUYt\JޏhI{?nKI%wtŠN^z%feHB}p^.tzܓ:=.~(s_r Ŀ5(l֎r"$dϴ1L)e"fhk˲ƈT ؗ/,yrhs(SV=:c5דyE/X't9?@k}mӷbG{CRQڔ3 x&ќUL c&)o -8e w yaN0tpI-ahL[ʭς$lșr6)R{f12usHɡ%|)bMm H>\V[[#;nf+[(lɢflbd'J=݄a[PFRfRv !TbkK@<<"vUt{<;b"`|(B0R &Њ3-HT!oL)"y0@ I' Wo~3I0JIC a0$3yj]Px"$owخ׻MnB=+( m#If O ^*!q=T(\. Z*KƧHԎ'Yx&!;Qւe=eW+?{6W!03uv`N.ab,itq'el"#XU/[y38nP0&u4]sCe9-sX9/&f4iW)ĿZ;Ǹ-W$<3+Q7rI'y|yE8ϗ3|qu4Y/жMB/ ,s2֜F?,4Fq441~ypO %y=-6,ϓߒ 7}{ [qcLd"/7Z&om믯M5V{Lk (ӱ*YJz9~[³pFmifelы猰4Bƹi|b&_S'+\AnJC>Vi"P7ڝ観ߴfpac,ť%鷙|&mٕ][(9=lO=*[D.]̕᎝kUBfT1*B e.;/U啸[?=~TEj1KArH#NMCoʡG)FE1f7˗ʼk q€nG?|2á/mi+d e1y鼙Iv律^P'":D.#M.yZ*#6|Sv"ިU3_cbYu;wKCm'MV\s.454n^cW p-d4>5~Ɓ֐ow A]>\j8`輎3H,SD'?)<9R'ܡp|%h(t3*!*[Vuźznf531] 1-'#M<C<bU{ c=b0?d cs}z!x~GW<^JW`:^6{צY뀏,ژա.w[Ga*o#l^Sm7Ցt˃V3(Zή7wQ>& N7 WOnox6N{[2"zt"}* KAA.g'(|w% M5zaz| 69*ٰ_+~.~Qv4*\_=`QQbY,gyUTT,+J~6Y׳u'Jcim`fK,˽"L= 62\\+C'Pu\J(qeW֍e/QeVf{4w<|\3xLEr9𢚫{8C2KBd"^̦كҥكF=j߹5YaYc\bEb@>'BJhi!!ha8·Uծ]5RZӯ}Z,My Nj)椷L8LfL)K" O˝SUe}v>nuT m sX8b"IrWB?(g9}#gѩY,@GEFl%0R;RX@Clt0ClkH(ClTkUׇDl?>!Y: \`CY@̄+Tkhq*-Q=W0j̜5^ 84jP帴GOrmJ! ᝈL$O_^>f(9@&b+ˉ.NbT֙A,L\:bLpuLjx4@QP g`֙Q6]4 FhLpm8sv `P-FLkb i Ί(ȕ$^%5kN{\= Ԅ" £̾prWVvWk=WJC3#u8B6 LwWRzp%vz1 з"_m'XWDJ$]SI;f!ZJoSF  kʃ=/vj:PqubLh6++T(B:WG+Ό! [Mʵ8ɮ Uv]. ]*1֘`ʈDŽIU\~}\)6 Ԫp/!hw\ǫRL`N QU A"!6ClT)E?>!Q +)=vjm[AQJX8ߣ{XD9 B)Kvnŵ04 L` (I0 j]4ږOiͭ! A `pj:PeÔNC+˸ WVw~*m?>F\Y!i wjEP썎\^\j%XCh'xEZqv*\Xqu߬rP * P]:B\1ePB+kCB]#BYHBF+}6m;vW1xp%Tᨵ̬J_--9ZUW[̮m]\Fr+~ӏV;uSWѾ{ݻwo Os9),yT$NƎ69%<\*p_*MLj ^ɺWޖ,z 4r  ennpk0&;rT?L z6I=Vx9=_ݺZOhV9HH?xo2[nή})RE50ցTG r\ IvU0l2: M77?c|7UuŪ)#Mfxx[hG9˵"$+R(sB&G9Zk kRN8iJ:7ΠssdMK1 18sO`)ϕdi! NM{W\[.*( 2 De8W{UL#z?+]Xh,㧲Qj*p4p=Y$l|o\t9Cns7[*)+ɗJd?೛/vu<6v4LשU9b7D]b4?YvG{WA4]{r6^\:^ ⓹Ϡ5?j_Zge*綔7 y~Tֺo#.xU./<33}j&ʽS:Zp&SV̰,s'£r]{0`8NjV -6-ͽYZ0s^iS]x\ TCDey./HNsH. Zr ]TT gAυ {s7i𛓛c1fzUu4y[q{4X۷+q[5ܒ%fϿr@e!H&έ'2O9Mg$/P,#iԎ{H( 1 槎 eTor.c" ^0gqS0&u4]sCe9-sX9/&wBPZ^*6ۄ+wX@wzk`,\III3?Oi4u-Юi9M\Imr1OA]nH"B? \ݽ#X,iDَ'c ܓ]5)ɒ#J܊ `dYdWUׯz2QY>Qb4qfTuNH璻wn"wU&*އq:0iMMB48\MLnfy:uasf` Vͅl0VmFWa?8y͟+- xTirZ|^x<߯WQ7ǡ0 _nүdﮮCzx鹴%1yZOA-p4(|5'\H)RϿt0|6\f9.Hvcf^Suk&  ]B$ݘh +ϟ7?pbt8Woz=(I~z nmy˱E@iA|(DBrY,M+o:zZ)~ƵHwX;UYUQەaPb kf>?~W4}ƮU4#,C']eNF]q+R 'OߌcQXt }9PQ:TKB/}afp4\>?%\UΡ d:M3/>W=>pCfY8,ތi06W7@ "RgΖXC0ߞ0y/ۜG|hEhwz[oqŪn+{E x:1Mw| ʼnIXxqȵ{fjOmNKΆnEf@iَhWB{:60]Yհ.Gudnw{po729҃SClmf47tvQ*>úw9FڏO/XL)hiq(~ryϟ4G.rZʄ2a^!XGX5u[9x&yq:#f/Y?1fkˠ!ACEd@3 &uD%ԖhQjc<\&zG mC3gc,G[ku~W{=+<<1F: lp6HFY@SJx->c}9 r[w+ {zgxpsڥ)2EP{|ϊbnȵ{*ۆ͸Ņhtv' ~&m4eO<+Z77b^mh7l/'z*KKsQw X糵N"R=QU׋'*׶+ꋿG%\jmU0k29{iʒYqR^ L(zP=V…hU(ZWES#) IGH#\A'SM%sAssLG eZ30{-#°hj i̕v.ŷY(?RRGe6!D >2J#U!aN> , *^CR$ jxT9NG> ̂HB9UaaR"2 j6r jx`T9f)WnK7v%3:9na*YzAy  b[ܛք1UK9f2G]rt0GPJfևٯkP N+1IfHzk]/¼U*\:9rESoHj^2'B$[XC+`^: )<Ĥg  i$53Gc5b6rkĺem.:qɱzQd֋׋^YiK X#v()5>b*ork?'Pa{ 5E 9r7Y,?kpOя]v\0,1&W/#?^j[0HA`QKT2oQ0%'XڇA~a+uՙ0#I %7Dx@6z6cQN)41D.mʥpeOV5QvB|D!8w_\.6k8_Er'2'F7!cΜAyH-c:$KiDE6"c1VYE 6-As@eFG$;A Ut(J6rKV;d2-wמ|>9혢AGO,uh*;\TnUۅ},N f/dR; H 0؎(O6"K;ô;igGs?Z| GA@1) J| Ei SĬ*3O-)H/o&Q, A2 Qxc6e)A@-`1F>J6r+m!ˑZzydm-f.WYrs#1^t}²LE)g0P (jmscA8cFŹL* *0 c KATBH"Q+Nz:SEPGnǮ[¢saqiDJA)JV, y`MDrI42@JF/tJ04`0I>} /Twc /ķw'{|owH=Zޤz" *!Tqּ+#3-p+%:WRW\)$Kk5vxﴴޜ:E0`tRpRp'J;(kt::\ LHtDVѩd ) E&<"7]a%Gyuy_j=ϥ! 5+d<50cӣc bH !$qf> ,Rg2D'Hkz4CRA(oTIhX^/g#gUZXufұfSadM[bxE;.6sMĥ[zDؕolɼF:pk)r6(^oͦ@w>(.p:@7?~Y&)`8݃K$ =g Bޚ[Qa|vIg_|U$aL.AR`E(F`{h~UY.U]䝀D)iVHpIMƴ2Rw\"s| ;<ʙ,yyiy1Yo(uBQ*Q2 kDTw7wl~%aSkNvWo߹QaT[οȷ9@K?Ƽ53.Kb ?"۾NTV^ G|g%IWvZ͓{J^ Aygj*øhDqy92t*Pi"aA9+뤪v\KIUܺH} ZmT%bZ*Xbkô҄] F;`% ɕ>MBZۥ;^ae^M@cvޯK~K}#(&`@)Bg-#'o$Ig^GIC_Gڅo,NCTJhiMVxYA{:%*EKZy ]TQkWW_>i;iMh@[Ll L^޺ޖٰ6xwށD=F5ݽgT{K%bN@;+9x5jwgC4clEN +j0#IxU䕢:腳p$QK/B7,U=&w)~}TëŻZK[ڶ kzfޔj,mIs b _*J_ BYP*f0 ]Scz(>nқd3ɀ/O6Z 2ϵ {cL:%4\'W\&.F \h8<89QȤ@|>hKS$vg 6ʎ ~U+>ê&kx@K?eNLLYFu|_ſu`4ڿNa[㓵{l N49-~] Xl9!RwK:&_/,=/ &g9DP(C%ֻ"ZqNysq7H/l96(j=u!ѻHv6~1n`,98 |J7t!^[N>śDRQOoOOygp.Ohn;ݒg,5=^qt_Lŭ.i ~MT^y[_xu=8lU0 f^KG.>NǕ帮Gp6:؋8x)Yגfk[.56#.waC k~ppq3e@$k[eV\ھ cx|9Uq,@8=Z _P+U&~>?^.GX*:sݫ|WoޝSfqv+xF`+ [ҭ!{sw֒jNZq=/ +us6 &%NJG4Sa rQ1=a-΀MgDiB]4 ͛F<4%j- ڥj]^?lW+\e)İx1^z٠>xI zp' GK/בfM1G#%.RR**(l1*<+DŽӑmlKY?[dp|{pmr3`!pP)"3sWqLAJ+I*ZfEKV{/s0~gaug:O<Ֆwmv[va|NB ~ڋm;0F{Īm0n;jfR[)p̥ զ_U]#,VgDWX( ]\IShm;]!J1ҕX]`]!\#r+ ?m+D2+Jl r ]!ZzeQ*WCWjéW=-tt $pjVU3eʠj@WN=5z#l u|+D"]!]&hHNV:\ ъT#+n"#B&\ +R+֊Ib"XO{1v,*hiAj|VŬ qmw_O聙ۼJ(!D%"ت_+GXMr.$V/~ r:c E2DkZp(9,HdFtʆRjNWَ Dڌ [ ]\JQ-em+D:%AUN6;X>tpͅm+DUt7z#D NWKCLW*rfhŁf(ҥЕS,#Rl **s+D+[OWR󎮎fʆ.c4Bf]!]q9ɂΆ+D)GIWQ}}ZaApxUzE1 y]k_V~{Ō5q(mWg"ZkΡqBmǡGȡ)xfDWIGN: -7m+D:Q ZC$!u![1lN4 9G3G\BӈhiZsF(͈̇5t( ۜ0C]8!ZzS,ײE{p]!|ADZ "J~=te6z#6LšU s󯣫F(ii@WN=\R]!`c+o soOWuߖGBt( cZ $Bt(~ "0; ]v\;lLѫ^ɒՌwK6˲/-,hriwF KKfts[T'kxtWam|MLFË8*gQ'k>7xtI:QwWx.D+xۭvHRg8Y3MF#0WX֤xFM}D.c44(hiZI-ʈb9!\GU+DMGWGHW0\`{,|R ]!ZzBS6+eЀ2pY6=VT1ҕUj]`t6tp< l}4\yUte7z# ۮ=t s6C)ZFW]َ:TsIFtʆ1=Xf(i%lN&\ ѲKWR؎R ++˳+D+E Q`GW{+a6# lADZOWRwc+ [` O:%(z7MfD M#\4(YGHRbhFteBWPϛ4]#]i+84B*}nY{{BFttuteYbjFes +DdGW_ ]S#hu#f #p|"˗/I1~;ȹ?H?_ah/ ;G9 G~1.ɿ|7~E v;ry}xQ5wgzsqd9MޓÿoC;b )r^."5m?}ğZJ*>./{ݾv>j7C*ެj>|誇WO~YZ>9U6O̯ef5xw?fg1o}"~!Nbw1iyWm݋r9jBZfIN  [k\Mh+;]M>'V妳x}dfKx*0μo}o!dgFߚsP\ݹR>tV!dl_y%4qSDjY;t_;35bz4*KfV>cឌHf g}ՙ!6eU6\6Pv|wGk";asX* Bɐ !siXhT\{SwEaQG#Bg`d(wz h7msyƥz*E,Fڋ`2<%SfcΔo?C4 nާ/j7 9J Vݦtft="fNqI<,I[Ú%-‘QhhM1Ơ-cAzQ<mt&$dTCK @mT/  eڌKAk IU bKT uQ]'cם 5PCY<` &`љJwb6tuGN@F(ςфj\!c BFoQ xi,v,*50:-!x:wJ͠j`o]\?1d a.! Z B@ rNw1n2tk&@AgEGiB?%` 5 ࠤ`g"U*QV rљlڃ?J  ڛj,D\єfA*ZtfIwB=@w/:=~ < Au[u3Zq*BYq/ʺ@J @[w%fQyŜ8k0@!1}AY8P,,Zr $F=(aLl! }b}F<Ό>tOqr>,f䥪*+fQR jrBP1s9;_a2TήכvH<hy[_ e=D ncj K8l;DqUM(]a_{d @+uXFS!yG a>X h`!3{D"[炅 2YЩ֏oT>/Ѵ\E9ҶAMRB(NEk}r~Xw}4.\t\FeQ}+x-F1w`:pNn~ ЋEuD! %y.h WH9v,10l5c0 l3"=ыv AnK PmawH u *+fݹ`iܢ1Ê޲ICy@4X86SDEȚ=R\k6nc:LeҪq4 VCwBõjg0qQc]|H4, -bl+'J[ "~g 3lӰ Pd4@[wi(Zk7nUՀ{9AEgLmU٤@5 ~+fգKPE`jrU7Zv|A +χ.ө >i{v}zq2agQ@@]wP7ם5̂qup0b˿17,V}Z9&mzashY8f2Șv rrƘj֋Ҍ٤㸆톢DlwP*\͡ x9Cn$Thsy>/E]ps(w[Щz u %, 2P A-v}q@ nw#V&5զ95A'7 #wth^䍡 +U C GuՈAÕN WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\J WbÕp%+1\jQN#_V~<;zǨ.vð-|]Z락aUF=k3݈h955a,=c+ v4p5uGW#`:\(*QTx9=nvWkh|6//#?G"c8VZ]#Qqn a_~z7#:βGm>إAjMԾ^E\wYM3YUM%Fîy"U_e)BjqOk{-(< տbgfHLe^ȥU}_=*e~sRK!OiN7O䬞2M;cj&r.@(*hxOTT mg CG#2d1}bx$'FhbԈҋ%3XBA\pM::jDշWޅ;ᅰ~x6d0 gOVrvVn`ջr'?{qu a|@YrAI AAD1ψ0E\JRki)"7̙יsZF_4hC,d]ڊjf85I4 #L.MiD+N~QJ}'HӚ !yJ*HB7p%!NWUS+,[nCWW5fY pCY]=Egl]`dc ƨ+Dԩ-y:UXMP'ٽ^p ǥ~k]R:tgzhS +•)thGb?t(tńIvB6NW /gzt%ި ZU)th-=uƒkKttJHݿAtԍ+kHS *s+@)9+)$ʂc]!b5-޼"]i- +Lisp \BB:]!JthmYfDi1h!1-3NWR3]=A$m]YM5.M+D+ɩT)?]|Ǧ/:>Ҍ~(͉-`{;Ch վ\NlS rut(8+fcAtj ]!\uh>u3]=AVќʘ.5/Z4)UTmW-tSb/2> iMBM.jJL@.j.QW[}h p M1-?yQ=OrDWX4hjB)ҕնޜTZPޢK(H%r?mځ{a4i(o M#c]J}HӚ(BE 51-% Q֖rЕ 2_7$M+DkNMgz:te|JWX6g6gh-9uu x߱yPH3{0ZƏCWT'AWLWzF\6M+DkȩLWO61tp9M+D4gztřAt9e+{C+gcIҕ0jUޤB{y QKe id2 D6Hi8k%ޱCNA(ڣ~u%AW{.IJ\Y. oM~z7lpOsB{k1̹%H _&>KhQz_ލ1=}r;,M^eI gp$BFJ΋zήh:w\n$%`L׮ڙD}E%]".fov۸ktڧwNcI's@!;{zr݋?WRD}a2H3W(J%+|1/c? ܯ @0~Y4ЯU(E?c=ro~OqS*M/)l ^t^W9^侈N+MW_*&T'j|փHu%(雗T/A/b ^1v:Re(E0&P'dNt rOHd,+ƈ$3*/p.wnq aue*b èmR@IեZdBBL";MfZs&5 S1S)R e' lL-Gu2-N{՚p" 0?;&6{TN3 HWÖԪOsf=vi]rO' nX#(0czW\Ǎwb ZT)nR)S]VvVWCr! nUzn+ do+uàݤC/Hr{[ !3'",A*$5.V'!q6{ &aֈTr.Xͤ׃lbg28㏙9\ܧm!C϶[H,O0q݅e}4Tf Bp0pXQt[? Dt9JTL GV ށz%%%F4ؠm} TAۭ?j]u\mo gnӤurrXt0iU:~2t1Ѯ|})A@,ts<תxT'?ϊ(:{'PIk펁'Mp+ OBN / ~{yv }'jBْ;aa\ K4YM9&.狞ЧLI"o.fVR:,Mg 5k֯:ٲg٪q{Yi(um zB 2N+([/Ӡ5p8P1!bZ >ZӺyv0|?WMs!?(-n4\! ?]o"7e[p;ut@Z+?UU ~jr[ zGzrx/WB}(b";vg".[c!l͉*%I}qkt܏O ޟ}1ЮΰCYbD汰᧘QynsL1Zs!2=`Ew{8,@\7W8خ,߿;R̤2m۔.ZF[@+5:Xj- TVMc'ߨBYO]qǑe%:biEB%ɪ"٠pYQ&3K}R]OE^>30i%c"<32@W HJfT%E yL> Ab6>罥B$239RrgWՁ:ZOQ3Ƃd!J:)F[JQ]xstLRigyg7Ww 2BZAa8}zy] ˲Zt];8(\4L}ہ;88URjpRi§ף.7:OE:]I!u۲~bP&l~1ezݚҒC$ hSe b.}P-ZTvJtp4Qfқl3ـ/#:Z2ϵO{c:g5tPGQ>piԚ(c$rpqs.I|ԖhIv lzB3^y$lm"j:/a w"uQWΆjY=EZ!mpFuD{LjVf"A#JM4Ij}KeE6hR9!Z[E_a{bm*UbZKІitֆ  MQBR\bBݿ$ 4HӍNcw]ޑL9MPI"fn,4 `[a6YWGq O[, 2szaAUr^ënR|_r)J=#b+?VEX$NֈPICz'M:dg`t:d7j%?\ŊB {G6kj%] 050Jb/}SS2Lyk98A(2ȌB7g@S3zrLH6c;RTr8p2L ɋ ᆃIXP8]:SA+I+3wmm$IW ~uy0 tcgziFfFmYԐmM?k`l#ŋ.(*iQVu6]U,FUf<!D6FB&䵉5츧3* 3T l(s8 '};QC f@y$ڎ)&}: hI\-S)L&0r҈03,PkTe _o8^w"Woc0UhYcmV(2xy* )R #1Lއ>8DU(:;Gr0f-cU5!(Җ> K<gcwd"shN+ĹG]D'@YM>e"KE_z̅h~U5Cȃ" +lIDt\=su4>sM MKg JJc"Hr`]&LLb`^;LBu1T!:gsBueW"B a)zu$7T1Yj@.S O˔1'NDJd!cc2>#ڳN*jkkiqs0hؽx mZ#5c\N#m45?=oKyI0=rd/k^6l~)p;bD+[0ƙշwnD -vsˑWU:Y˭RedL!*jh?"º}SZ30e_uDmR5 *Qھ;x2yVYZ/Ω69,jL(%#]R*f]'KeA D ~Ȑgr8[ܝ\t! A_q5]_KkZ'inn9SVnyҧ= 񟵊.m21pa&{y0zeXiB E6ɉa/Pr?`AՉ8mKP ?,HaG##mlU`l̀`%'<`-!)'ɔ,rVI \xHۋTn`8;7M,CiJ|:ڮT nCtK7KTjC?pC~<e/eV !JMDI'gК)b2(.<=|kd`'@!^`/ b =!Nq @B-Y4 O)zH'eҜwkkLl,Y9Hnv gB$pHWF"#/Gw7c )^6-6G áCw u/gwpOa{!+K5@(p[ЉǷL34ƹ6Qztpr2b٨1d}QCw{EqP|U5,lԆB tBw/כ_ڻ2e0f_Yf/ZXƟ2+tr5wO.:3k.u?|ގz>fMRoj%5f̼}ȣdqnu3)+o7j#kOiq!*SZсH4nt?<+wN~*ѿ|RXwܸ9^v:ˮf9sJY3$a4t!HG_KY[ggexN:U-eEiO@1mpi*U#NȪ5/Dn_zoD:pu!*bq΍b`wӮsafBwgmrBջ\0#y@Fp!\ԙG֣QT0UCb Gο=P xZ,[v 7"D\FC>B%dr/dM .EhЫze=ƎUcjT(R.mן'@͔phJtjh!9CC8K#_ڏ8)_-aVDNY+^O0x/%Vtb4?ʝ ˫w=ofٻQ* ^t͆B'Nnt-S2䈠LOAy|3<|3h;сrA;0Ι܏T(3kUmp I(5TɛٺXڗ˲zi"-z[:z V8=物3eD}J{0́Rv])r|,Maݥvupz+&1b!Ӗj>i3Y膀]U\6Wc~>6;ttC- qr2^Fk-< %[b;XZdiI.WxW`ʇo54m?FMQۈѶP x.Q`e=2fPZ'(K9iL0˶Ϡdɸ^ZІc*;"DpE="Deb\)"3o3'x1x36 SQM-B!G<S1K01&EJ,#: 2FŔ4XUH@-Y/7ҀS!`'T ɝSD ZReE&݊Ji;!'|gBi6Ov44&gMPN84'g]Cqq-ѭ<{S[r T\)gB łLlV Εf*N,H[[8r@# L|}\*rnd\ j#c5qhXVfNXXr+cpi˄y7pb'ԣ)M/iv2dx:z]j,"gK)b.42ixTN*t2+z QHQ`dT,UօKVj)ś03*#v5qF8>5PPyM=j@b&Xd^d@ZyTι$,1ΐV))sNy uRۉn",C8&%#Pdp$!B9hCĹ;~WPEeD"xی7"ZiW qi͒H b0V% ꔎHLu8cƒyaP+}6KY Dhp+ѥDPVU1n _MPG!:͒]qTE9​T>-:󔂳Kig Ftٍds(@[ \<<6;vC[wQAk^[#>m'!` n`?>ygҗz|_^9P Pi<2U/|C8X2G%o^LG~ ӿgztq9wD?Oͼgdⵊefcs1>jAOƟ'MojwzȤO^s_°T١O^_9'F~?*LVjnY q[-rr`m?GҺ:ol/eچ @ɤ(!!TRgX3K] ޠxx@I)sDK$nɵ@J[M5F;x<gsOV_7fb=R x%n>U^CƪEǻS~3}|dls(8[G hηx>`@SID]J t)8z,j":-ulZE76` s'7;=г#.2ke)MTrC=dSFᆐKEK!:43ס9:4Bʎէb0y/EyY hŧ#O?4Ɩ!H&U #WsճW|nmC966J4TkjkpEV`pg*.ۖjR1XZ,,^u%Y̓Ѭ~_6~'x?Ǘ?:2\qֲ(+vA kjعP7 ƶ3m0#[>/A%lJڦvЕ͋'ē$.>hwѕUdhdv lJ1϶E8uiKb&{un+ǖp}eyf'e(M]Rh\h:-Cnm ԝ,2#K$6Z69itN< QEpX6D )skZ,tbi:St2`m]gMb(FDM(uε\T1 EbyOע[O] q28;M(I [(+Po7EqTݐ& lW[Yږ $ےiZZcaVڋƶF<;]G<^ Y_Šj:tYL҆\5FѰ7:*ˈNwl06b(kpdƊ"0VٸPyBCB֙adpvgL^Q>q*%H.P5]MeK&%q2SSJ{w?,sU{*v -wC,C[FZ;(:r6&3ۋ~aK!TSr"غ`\CU-jRjc2cHd:);.t u6 5N],ŋ,1;l 7V7\fmED`M*o2Ie0=JZk (*qJqؔC% ] ǝ1VW>2;&Ӑxv1A?Z%EIn j d%P Ml3nn? rOBuaMUTm` 5Vyg̋u uI0$3]ǵx }W,&~j56X#nڮ3ܕ%ʶjr8)jPl R J`FPMi˯3 O6XCuhyӏ~]|ܧ~V BԆEٔXkZ~EWMM)xY߮׿zӯU>{gFLJ^Y>~a^|%.v]} @W⫁lj2.~{6 5BINzq(caVE'ԣ^[&Rh][rNy}{i*R`CV?[/~x/6ʳqb$F vfʸ^xK̴?~X& `M5VSљh;3K*CcAF/z򲩮~Kug*]?^*䐞ovX{k`=ϸeïFy7Wz(}۾[r-/W]/~lfJ7 k{a7Vƣuʨz[|Bl[毋_} VIkf12^iN}#N?ȻqzwuUȝ)^ ,9\6]ˮ1%q}jq'eߵ"Ůhs#OEW|̅:])3]!]9lFt%Е ])ZJQ B̈Ђ9x}Cǡu;LWGHW>j-,-,;wt*v,o71h Dz̆KC!u`f'(ipJ׏~½?t{]j?QERnn`W9˛emYu.nv*_~C#.NijVÛ!ep".?ߜȧ,QeUo]szT >;9F$|pBS5?Y~wO{zmE1:lQ#(;[%;|60Sn ^ '(AnP&M"q^.d(ӻLUf^BO[ya͋6X_Yo Tc>y L.e'(I,YY>]aStp!SOW2ŽWFLW86ȺktFUꩢMѕ ])\Ά}nO{Rtut1#d RЕ=T q(CʱѕXW !d+E`+s.µ1[:HV8b{bu8{!ZX7:8:#iLiL.4h}4(S+2t4h]Nt%!#Rl-%#(3]!]Ήѕ&ǺRuh}t(S;#1Ջ[BNt%9#pBW6Pt(i{=tw}wJQjtu@NV|`{2L:RiE4ttbNV벡+ 86&#(93]]#"] `vXW fІJQtuw$s]C]*zr+̔ ]Y$sA8g:B8#|~R ])ZJQl]#]9یJG3pr+E{JP%`ܸI1@̧}è9 u<j#qcǃQB֓x1 ,Dg)رdwA\NASV.8GbL ѕ6 ])R+EIsD([;*B sνG7h}!a+iZ;д M+Zƍ0t+h+볱*mL V4g#] ] `oqƺR(#]1jAlJ *JQ|=t;O EIj` qp]=mZ:5%$Gwm͍r=K33d$هI[Fg[eY1!21F25ٍn|nt7dcE=p Xi0pe%$2ԲÕo'D: \!&!`ʐ+p(p Dzp" +C0!! jyP)hWWT,B4`ʐ`(pepe|pU?w4 JτS6IQHV!r#Wkj}iC%_!Ls)&pʐQK@R)#\BRF0 'ÐY(pep &p|1p@3#p4\9,^8fH|&nJ21;G*{v@ lcDLWpY0peE<2w2TRnK+C.   =k4p!"W]!WZWJW WzJ7}>Ϙynj vlJ.B9‰1 ݱN"/1}{M(~SⰠ'ؚ`e0.!K.]bSxD|,P0peȕD'`PAW\RR6&X@\irx-Wk ?Ò[CCי7$%Xzv<'☤smW v.< ږHRgțIpHAў,ct f\%tuLDO,Bu;K\[n@ښ`^IbiDoѴm8MӲtUHBW<Zk@Ni7#L SɁQ"D]7?Ӫn}a:Y-/h~C-6F.BbIH8Mɛa ]q, ݴ %$Yb_Am!s8&1og!ȢA/ -T& %C/L"(E!Z.B@~NjĉPL 2ѰX̡ gqb {C/@zd4!,B1즧$$d+‰IVF3 Zg{̎4'cp KÒXX̍/q8!P ϊQd@ys!iZ`L4c<(8/T處(Ol>h&A,ʞqОX0y70ݰxT/@ rg Fq2eyF0d 1԰ޝTeFn|pvc~ n*a(NΡ8N*'eG5D6T^x0`H+ }d}Q+fȹJ lr}D'@^Xx1= D||smlU۞gq𯄳sQ&l1<ƉVh׭  ~ýmLҩVDhpfJ2Ok)]ǔOqВ@fzueY(Sc:2ᒒ%^ <ǁtЁ!|TRڞ_ap_ ʄf~fvkGf0Wf6"LDeb՜`L;G;= 2'q2yOөh2qkˢ-WuVKrEeF>W~\to@8Nff8z.{ʇ9CŇzd.D#hZhw$-QVJ.E"R mgj=&V˲\L]x.$Ts)<(|*XUa<1!ZsJwǭ\O,9.5LI#;=Lo 4֨ Le1Q&s2ht t y#8Ɨ@ښU-mlƇP ?s$[{z4y>WLD>,0)|6槿H;؝8"k~]}$Po"K/<5ˮ-vh-\pFo?ec 9祐&Xr >=mu@Vbp"1xEYl73vCiwOCPUAHI] K[%vSu*'&׋)ywBo}Kq.vsE &}He\Ժ/4`NL#؁D@ŕM&e*AkZ}.0kwo/Y zR[oKv5? _2}ZH'Ņ-?ΗV7QbU4IF(HRHiDQ `>kk|6|}ؚ򓐾0,ө`zift^z\)^YVwI>Z4C/y]O(VkkT\&l5sR΢Q"o&xMB@>!̿VKũ}C26G:R9Y}ݕ4Xs9'"[Prrn.p"QD(EJ:!v\ / am>8fM[-Mh $YcaXzqPO<΄0 P+qG0rCczn3z0IɹxK۹TrӲ-ٔ)gʪf0أ'm3>o})e#]ݺbjߌbI"k'lYz](tu%J]Scؾ-T0~ G1]D˶W4teFnuue+G5#HתP `B".hCH/ڞ魋I1`ɣu2_ :P,/󹉒U`\ʶ_j=0gY4.~w;cVMi0V9l2p,%ij'v>{v~d{.e~BN]vYx-~>S˘ npl 6(n &r"iwQf qPi\ri;=a $z׷N)lB1- aVoЙ9#FƳ4(R;]ؠ[梋Ӿ=Da Y0{;YR^(sۨT(]ͪK=ak\HGޒlyN'|ŠwG5?1eZCldyN>~#3ǣ{c!igH {HocXQL4|l1\n3h%fF}mmѤ(6Z}XO.nZ4U5xe\Nem2܅6(JY׷ F2ywGHd+`Gi#bN_>^;N#%dhB*8v1Zuԝ+)[r <^(Mv'>-o?['s T ((L(R G˸5Շ-]^Ss7exRhhAg=HI aK4&ld<+,8S+nD)QTvi"XӠ~Ij>N\zSuɽ5F #jm1u&?1[30­ Rޗy+~< "5Q]>Yc 8kBp,Uq/bq{SRw"͵URu$ec:>l"ؒ&%cU*~^l-z$ $ J$vi*hl<-B ߜ:r$C\$i($xj$͈;;4aR/"9Hm.,!Di"qϓ9%OH+pW@5j5p Ǥ;pϥHFyDY YPm+M)EŅ?lhCHBǝ?!Y*mp%(::AtI$3|aUؠ4,*FtKqm t-B brf(&$! NqģP(Ib9B]>67CZCw _4U; !9A1Nm:Pt`yDSE,CpHSmkꌢ 6q4aUdM(`l1!_nZ|m?\\7F~RgI[,%5L8-7#rJ`o^r"DWB&W]x8Dz0E*Bms|iNN' A1}9z&ًۉ UP:˽q.< m20N,|NŤwu}#ZL\(_q0էᗻ|X RV; fC-Whnd4{b\iڕ8:.:͈wZH8 %L;%$m+dwc4s{O+1{| IQ(qH2`B)Y{P~ÿ́;,'q ΦaQ2U9TAP),+zqXBzQ!]cyt/ טd?tj.v$$Ѡ:QbI3 Շw>VzoBEجkݣ1nwmry0k$VX:(g. X{C\.վN}0 5+;2wV:`al XcBK)Œ(k-N垔`OT;J{!Ld_?՟'x!RS˧,OPϏFیY)ܡ@?^C?VP]RM J9X;/&X rnށېKd I=dyꋾA&w 9w:7E~ a?E3*h\U0Hr ci.UiMlcA-{~_C0&6YkD}E> 2.H >ds-SAci3Ks-yZJ[tc<t>"Cwin/?B9K@#b *z4 $β>˗bdItxH #qqe۹HxPBH$"{ Tт נ3e/UZn! q}ƐU}f.RQQG\gi \e,5H|rWfpUQZZ )O*L "=>k+QY ­agr}K|0,3g! MvcjYnWK#ƻt ,=lg#WBxvPWpN@^a$:5MsubIzPȚ">%Ѻֲafс΢ Rvp1p$%=NsI$OVIeh DvJ{!n.xLPi2Wt$.~A63*ΆH(3Tښ.4+o8oYmH(a_W Q߬YTxPRJW}c0PuYN|<ÒŇ%U&E2ڌmY8V4 &x55$auA VߓN[UB ͆9J*'0K}Ȩb7142>9(3x2|Lw@?ûNjQǚC;4':4}]hzَ䐂yXSwCTK.UɒUJ1}@BVDqi;T$pmJ0hd\2ǐJZ2P6|q+M.`2}\6)6=1芖sIps/ O<\~Jn5fpT5.Ne.ꉶ]~ (Q_VY6u+]Yj$FUʵ`V RA{Jʑ3zsQ_Pϴ\n^e꒜F]7[wcuPCC:<(w!es,E r߸ wWI!:ۄw|h^ǯl6?ˏiҿf_ج4;u 83]&aOv53O<[cd#[&?f67HPm/7O@Z)B￟~o VL)7zAx kcL?~~u)?.E/?f"$G5?,8df(/0XCX9wds$SQsw!tS\G5å1BbCL Y(nd:~ #Z( #2$tޢ6onl>2i?@R+6aMT6PY\ Ry{{OX7ʻv[Orso 8V(\mDSٸ]'3f ܽ݊ œnJ+RHx9v %FչZG TcgwIa&-:bv-8f_N1{J˶t3WE8Ū?O ,X,zFRLr ɎٽbҾΓ4iQt 9wF85#eQeYN9Z4M~#d2 $W(S.oJyLV 5/sK@eij}ӥ>`؄?6xe 3$4sZĩ|bcXa5JX^^~$׬4R,vGS= :"'gӮvIMEkl8LWE@߱`W9+ak# -ok'zK15fn9-äw}K~]*P4GmۦI_')B1i4XxHW ]zSuQ(azy^) t+v.: |wEc#bogOœRzZG1lMY4iH~1&3&niw-Bs2Y(h1؈|W(cdK|#Ebd7 P%E$J'2A x-0!%y7c/IhIQ hauFK zKߘoEb]"xBL(ʤ`2,4&uLFvrSɌ2I]5@5O$8h4eУ- /yq4`3[o> U`߷n֦n&i<-|Ɨ3BSSN5Ҕd -O9"ҪZ#^FbѦ#b<2F ;0l"A 4]I]v8:sf $m3d)E)Vҕy Fϗ'&C w+/|4)ǘ+aSklܝ|/ g:!$kO9CQ}w^ռ[b,?|ykx))J\ɮ΃&,3FB۔F،+/vRJRg?6Zp6mv!_r#C :BܜȮ{V`Tx 44@ AF$]Q$ܑ*p̓=4e$Uawi1V(F4=PAo hMD 3 k1@bnG]S'][JhE>Mn{ᚱh85xO<h@>NiFbYop}0^)S\ a.hL+^ȤW\0T]T]%zcHY79AXxIjl:)Yw*j hMAop5%%w8Su=Mਗ਼\2 GZ8\f " {SBjl|ue$vwPl}4",FowȞc^ND|~|G9uQ cqF+5/o; qw7=1<~zأǒM'wz0T, b՞8/+ ?wMpg}8T0YG4aj(&ΘuPZ!N) ;S]ʎuͱJ,\g3Y=%Ensu34bУ[Gic8"1˓?*<[VgH:2$sBk1pgN3%vDSZ3M4,%e0UHKNu"Ƙ0X, 0k[czez0+΁fJPC811 a9Ҙ P%It9)vxVIe:0&jh+p&Zx1(e0NTG\Ow(!㄂k NSN-r-VӀY 36e~6͖?uh!յjpaVZ4VZD!PɆQ{;|+{ȝ13I;8Ԥ=>ZE(q81\݂0rG_BE (WcȞI(Qɨ"4 ך(e!84OO;e.MS<2zff+MƝ"*_[d?p."1*l R2L*P~a"V![6` 8If%—@3a`O{O{ޖ#r2&MI>>1&CR(oU,۬KTi;C Q5g$,uD\qŀ4g1y!/Rh-~J. 7ou$ Ңdآ|]9kQWQ:r4y@[T7ֆ9(^aq:.Yt-W:`Lو =cw_wG28:,Fe}<5)U񳫖3=@Cwc )eu86*̄=BD)zD'bk!cP[Q h|8sr5t1l\5S=x1e 2χ:̚!.S)1m/}u4*D ^/nabD.xW4ݻLJ񯙔16δ?j }G]uh¥`1(اͲ9S!\w?5ōtF&n%R=~١ k(Ao$9sCpZ-Vdw [˴Ht G=]hQ Xåtq$ZiQ_6x39-aѹhdz-lcpqv<*˶sz:uQ)MF߾}75L4} 6èp)ݻgI q Vͫ;Mhz\KH:!$kbD \ Wh" h*#dgp=^X k{Nb+B~TK[>nv{,<< 撀~”q" Wgd5@ބ`-t; zpdzZHg󌑐pTIKlFx╗V;)%ZJ6tQR=6hi>"&u~KL# .,Au*Gﴡ`wu=&(`4˞ܯG'x>$5+2bĺ`Pdzv6>Ytkbʍ2L7%0N֕[iLnAw==es^ hwJSnl<zZ5f^Å.Qpu!wpgLPㄑ5z.eM4}|*T୽&f[ǣ_A=:}ݸ`OL'AIʩ9 rG$^?¾ 56(Y/4yr/9STጩv$L]O{<4@mMć?P53n6Lbxf\LcOJH\rK <\Q}eb(Ih/-;Qu[XL.'y:/;H[ᢘ1 -nh+UwJJj؏׿'PUĦQ;!M\CB\#rk AEDIiuS' <2f@˹^m/j`UOezI2Ar<]ݚpfԼpqgԼ~lW>qZnH!US*j#6&0X[g/Sp.*-: >J] #K4S\:>m,):k;۝cq:g1n(v,*fPM4`TPGoo :ײj'35L4Q{ƾJ@888+u^6e<5 b߇,^7"s]SISQl!F*dt|@X ioQI?eC,R5 ҚG(BBZ{ ]PT/?j1gk8?њbQVmBWB+{oO_e네*DMaܹHWtsQt2M8u/qz Cim3qK̲90k&S9eDu6g<Ό2Vom:_Υ᥼a1VnXXhV/[:عZ⻔(¦aLgy6Mtt==0pT;S]#^8SѰX5OD!2wZ&odZR(Is d|g9ǜ?01_}_?@8 _ʮƙ_ Vظh}OBZb;s`ƨJR 7j@djEHY)R2GJ (#Lu!c$\I1fLʼnh=Ej}"56~e?x\\K}ZW[sq O`$i4p$6#b4ׯ݊[ qyϭoVrv)m41k`@TzЌbAMc?cD9QT=jb]й_:^ZWLQ+Rdi\ܤB]+i8)tږh+8Yрw?mZ)(ˮ||5tıq"L-ԢzBL9jθ1C*ߛ 1>ȁ-Xz[ԁT )%lN6q]g/q o]Ny7%2L A/ 2tn2훴 ߔ )5Ol!BqGy |bT/F;}S2{M!CHMp!1'L;8N4U)SO;vp܂Hb a\,[ɑh-U2řg溟"9<c@?/ӫ(\]Vmg5ʚ4].GϮ(f,*_ҽFG?b"(qu$C{GGD)(Q[ 9]} ҀMFBiG|Ye~ۻt(ls+ R͑5ZH(r'[Y;AZXGޕc@VK^ " pxґkFC_G#p#s5`Q6SdZ߶ F2ch;aS`+jAQT& Ap7=yP̥-}Rԙ+E K}hVĒtw\ĨBt|ZPn o4W5iя!ǘRF^ py\3{P߇-(4Efb7DI)yFKE.,eGZӍɰw~cD/mxfnz#[n:3"}HĘÞ Z"KGTSGzm旕6PCSĢ_:&j{1PNC:0q?(c6g. ΄@)Œ(kt7 x\3q 3ŭv(tɘu2Wb֧ZIM`rIuS=էz` 2 ٝ /{ "ioM~]8lj_o1]uceIܛN#=Og@|:믗[}T [?+g%tyK]h>}) 8Qђ.~u`|zC]+bYglߔ,O'.z`[;{~N6$S<ٞY%87ZoiyT^W;K".A8$[r1Ɉ?Y=V>%Qb*X `Dq o1?p ©Wd[t%4l:kN=urIn-z??n1`ʪ!0r[ja궧JZ-KQ]w; JV~4D0B=RsVsR1ڷkeZ=%"uw1rmdGGʜ1R$8FR&ws3F&s87oPȔ^:30}J! $O=m.J4m;皟Qhq`j+BDch#py1_e$93)  ,$dtuBE 0vpk [KdDp 4"U[@3'[1f.F [.3#tc&DEz:J;k@ݗuAA N1#JN#>ؕBt.Iz8qn}=RȈ.T T5 TQQeH8jb$Vm0jfF4cֿ|H:&-BuXڦ J$8SVP8܎LIls43Z[ H7HК*ǬU0uoFʙUU}Het1_`tOK%%+-ICs*{LTԯr_rבTv?XRaZlP*KYsNlYZt x+N04tAX'4tnN*.|/^-*x/^҆Wpi6k0ǖ*q(5_F-HQ`γ~OFR%.HD9MHTDͤ]>|^EB[,k!{ $u$]Fp L=IHD]ڜ0%H6巃9e@`2_d`Ch6Eƥjy7ˡ>oQ(l):So7,i&Fp69/MϦ͘+d2 pz`! 3j`T֮wW?RH/#U;\WBކsi҃}wL 'gy)(N||\aHT"xKwq$KM9I(kv0=¾;I!ayYRuQ>M$`O `a{}@X2[.jAوj%~B_Ҽ$pR[>q0Xw?YjͥEF̿-}ZtT1vJaޙ7{7/](yFUݤih6p 1[e(~^fwgQH&6wU| %X'2]RbGbT;dTԻܦ6l kre#b t5[v&1kk48uRCn%[W.mTW)D;aʯo~%?ly#ݕ)*,F@dQdvk,4`0||e[ʡ'uѻ\ÜRze4)5z^çK  FrRΕlgn3P$yagf@i4u.ߝ*#Z-ХҮ_cM$Ms$-):P/3S8w(>p31b+fFgVX1()0Cj nW6ncG*A>oE<ќP"HjF;c SeP$ɵ;,@[qC>e8ʐ5Z ^ GY[1OUZQhZ3SsN54"#HL!s&yp.V?f-qwM+ȍ]~slZp1:Ċʅ'2C R,(a@B)aTcki?.xUv)R/+?wƀjXXZpORsS< cu!T0!`!082Q)֨Ɔ{` % U*j:$6+wwㅏV=(MO?0Oi\?<5ؽάlHl0<O}}~{K1ZFi#0_"8/2;kM/A? 翄_ʥ+Jnin4I'do/=六ԋDD>b1g \4yqF2 H%lMHj`TI'sp¢jX.BvtP(S_IUp"R{ϋqr$+Y198DVxlTAqd| up)#EY"48 KEz3^>F(̓{#YAO2=vX<}hSMѶ T4Iișby<35U] c8}mԒ6k&7Њ9N%$gs.t5 `0EQنЃf*8BzҷJ02 hT)ɇVI~>K)BNbfR` jahSCHH}x+bچC=څQơX02qދQxهӰőQM&T`HSYQՈƚ56Bb6\4Z BùW񒵆=j8 _gKq&`f쁒ߞ̶Wrd*LX-tFM Mtt:43PtKKXiAz}mې480}P}pa$QȊ*XҸr;)X2Zt لm7*-_>B o4<Ɔ0!3"gz`xOF-^s!&ƭ2raW4VʲA6|E+]xP\p:lo\+ag5zBB'&%oOl{ U 8,L+RTO6TwL=|-: pLh#ԄxCr,;`?`/e&RH M- S K nRM6&*i#!eͣTwg nq.Ʌ-V6:%rp}_ԁNm.;aG 5 šNQ,! /Т'=Oz!e:AQikޕPw~Uc}&}":!K<9[dpoudl*V]6xmra@ D {Z~8D 9ݜ9}x„E 󁔜. Tj֒Ɔ3`}r9m6dp{% έC'K=[h|eFRʡmv.-$.c , `=@-]T^y.S\Sq3m<2d982?)o-m}zv9.;F Y=gϧoml{}j} >Zm/ϗ_h-!PCFVJdr —x%bi#$' ٓUՆֶAX^6Ng[4S탢3FO;200R2^[ҸׅBhHHM6}җI%ȆꞰ͊uC]p.J{IA.$ IԪK,?9|ux.vpa kvp9auӷ__;^ ߼sOI 'ODd@$823AGR(}Gm~2dzCw7Mː6%;UpQoXJBI %89X [!5]6t6TNsN]jձu {u Oܦw,X,0ɾGZ7>R5e 7!$V&3mO[ z[{P(Rۿ#]w5(PڡY92ݢ\!:kd|D8\mwqs.&bOO?Oاrum4ӪӤ63t`“go-(CKܩbh 1 f %\e%Fl!Z6ų%ZN[Ur f:.zE2|*?4 N ͹vk_;-84a; u[OfsI7^'R0^!@:P P+\tb5hIcCA㩓5Уf44!%AJ=f5$ym4Ȍ-~roe^XpAْ&p$QW}v~Nɿ]n_yq -Eurgn4kudJkC]w+b7&&-!vj2V>e!ҁFky-_S|ŧ.rT,Gu:Qs:K N JLP% WW &EBh+X\dZt%ԩ +ݥin?j{T;4`*ma ʠWu%v%Li}2[$SŗR/9~lR~{}s}>"-}ܼ*IZg d20A~m6oD6&;@HA2uq jab5!X<\+XІo*CU7w;Zc/CjXېhL38V\]EkE 6x#<.pp ,AT3j], /wf5yI~QM~mHMfsb'{lvF6y_%7RTnRj?_7xOZ[~}*;0sY|ec˧翱4:ϒoМ88vm׊g[}>P['Yzbŕ_B]q6{@8W7gгAr8괝Mj4N'*$Z'%YU5Whe _8#_:+BO/{C:hO]Mjik|O|Ka~LE$3_8ȳu4?'Ÿ́9jl)^O'l ]}:88xl$E`rv孟ҚbuyF^;8nd>uIy(s~D2X]: mI$ɮJem- rSWBO]w7$L3m%=uta1ZZcJ3:,Ơ|7C{M_;黐Kݳ0ؘԗ:$I?CҏC\,KLyc++PI#1#؁֌F ]ѮhƮ^)FhĬo+P1tʛͶ#FъFyPrG#f4bR}Fpyc!5Fp'وUԚvEr#V)nGpSRtW8nԴԋWl85bJ6JUP4BЮ (Usv5j8gGh7Bu͑E!ݨfT/?~jxio\>"x!G10q%0bbj(*#TonBo4/тH.[k8 m鳕Ř]B|2tVXh?,Ժ*r8~8[ ?)?(!b0a13'bJI'2)^a$^wVdsN2LEx BDF;XJ,`uypX괂d;SE j,۰10Y|C%=U, KX1:|Fn<'4{s-Ă6a>I\s)E y9`-Ҋ&^ZE 7cVm.\x1pLkIE,(E9us No/{lpw<œis3:ɠm:XTBAdX][$E&z)O n\.{ wGݵn ̊gv:zR]eO392@D_m9lKTu(ZҦfbX&7zZ}j EHJJ;#Y]YWb>b Y{v&8vY&`#+A2<{ixe#ޡ(iSӫOD"k&}L=:EՊّUc B50c,n \5Tr6.l&vB!{118;ti nQQvPLRhRt}ax$Yl7= !\a]#)\)>Ԑ,-㖊xw}9-UǿnXQMXb, FWcgPE>M@@aTQaY\)N&hn8<ݦ-v[]\ 5B; t"ƞy`YNS0t߯Rl{)6,BՀV ]_Ui0Y 4i`PlZugġblqDŽClLBYp3vL˼iZmIA`\ /&M@J{ 8` }@z@!*)ñm`m]gln{]Ca'>=c Ι\EFTo Ȉ_SڄMo*uc2?.β8w;a/Q-UjjTKUR'9sLn- .V2fǤZ;} Gv4%l&[sɪ΁A|l`IQ  e0 s\NlbΞ"Xg[gw}ό.7 2k^ j.UȀ-Yۼ/Dp qP-+N$7mKu=G5s7|Y$"4bХ.,ѸI6nn `zBhRsuɾ gRCqym'kgl2tpӺ[ʹDloԋBu(⸂0-MI0#|b4C/zArsF-Yfe~0+1~<eյqRM_Phz) ;!"]ܠScF)$n:'SRzǖl 5:H2E2Zj? />a.6`xm8y%>hcփzZI'\=s5zQ8"rעEw'{-zVbt[ҦEcmPӦv|s3Sե!}OR b)z"lü)bj/>1~jQjͣK 떋W8{ZzqاE˅av||?ayz|sYg"֟nx?at|,6Y9iey "+PW^>\a $_^o- cB g;ָE$)CchV{]=C_pf[VJ!c=s6Ya},53 e.֌/ vOnjnwt 3w΅ձ|32q>'ɘW2[|±0f3rFﮂ*շGg}k}T9Ԫj}oI:۵DTzV2MMԚtҮ>9k%g f$ߗ|:OAEuJxbmCl#a^}wJ'.aKKj i%5.ZԷ̙ :$Yքj$` 6 ~+'IF>/2{d[_۸bZ<}ck)}vIkv=ƶ6iCd4l_r=A1Y+w|C>TQ>2Gy~`A/D4PQCya֗7k~-+! Љvcph-@eg&eRuMGٖJOŸS.ܹK{&[zv+},KSޟkula@1#r*s,ET̀1RԊJ*IʣC6N*ܜ$D,)?G B9U\d Q:\B!KԳlN蛯n9[zXg[+ҚoP.\BF}=wd0Kz{H^tpO+oZs7=CŰՒ]5,r4g;gb P|;J1(hl9 x#^[)^=! )F3 W6 ,4iMe. lL.>b-Ps䪡>#mVP;L ٺEa3Ue.J5rQR\j=Z72:+ v g}L>d3ତX!)#q,' uS|G[ݧIb}Kyְ'rxZPǬq)c'GIP耡JLDxYNF>hw@ˆdI -N @ -O`oʌ[)Fzāpl4 XJ-?!*o>m\fV#83OssnN!m~@ږ:vg8sƙs5Μ3tgҞcԁ~zX6"2*J -Q$r&MEZ\/#bD$(uMoކr3Έ8.McUﻐ/1?7Z!\JB($m!xP+ N-OCpmC0|>SDPH0V|۲v3cHO1wo5LQKΠNV'09eÏ[|v >JnCT eg&VxjV } pkTMq{ȟf_Hn[p']]s\7+S~*ATi&/[}gDvdTdrZVF-'꾷o5A$,N̮y 쁵e1J'l+`qmRPw|7'%ƄDZl?gG":N}GY>Ǭ}wa.]pcT[ ?O#mNk8l=Ojgχڥ 'g&W_nQ/mqQ'BRb8Q5g"ޅ7;Ks3c^RUJmQ@$WErQ?I5̡hz0Ӣv7%Y*rsb}}>7~@͉ᵻWz{5~>@'֍ϊFk1 IZN>Cbt%$t ]wEL`$H s Vs"=Ij\|bNyAqWO~b̒a-jw?􁰹¢vOYwEN).)ǷӢvϝY7&sv,~p/[~ 7`,jv"]KfhR>IP^nQ3v֍$gKb1' pvQ3vS{*̉g{PN"c,f^xùfv֍)U \S[~p₇Ԉ+Q-M)Tvj##bw yJ$1>PEZkrnMcWDΒ$X }:-{H+Bu(Zz %:+@QG!\=kG3޷iKkDZxK YwNIdHl mVFj1K )JB1IԬy X +ZQ+~A-JlԻ59 Ta4A5lԆ:)؈j6YAO}gFl]ۥlʑ>^xs݇J UIjiY] f5ÃmPM^Ĺ\MR0t,6%, [Il68G^$fcb믿\n׌ʻ8-f"٬RsOql-m4/e&@|#I B3\ )S;1À,,rJ 7/So.HM㷗llB+"^]+X~VCCt C/гts`TD%O0r/mds=c(FÉ?R{7YM$jKF^خ32@.`h"Tgm3䫄(aJTVi<dʗMzd+6{,(6q9i;b3p*v(ISɱ[wu;YC=Iv]4ڱ+ɾdJ?${T:Ih ",ʲ}IJ==jgaۄ9u(v F.y\nQ1, Nc3vr<7#Fe -j3|2ԎU6̙U s@!-joI/j̩xg@Z{0/FDZnQO,⽨vj/mA( `b$/ag?'j'\xP{Nn|F:<sEw0/ҢvG>QvK0VnجB:w`Hp*-j.Ͽo=yj|g)0/dfC[3Ї^6O/jvyQ0'f)#'M Lv̢v.j̩zᬲvM UzQ@_,=jga~]SJ`']{ۋo=ϝY7n8;jԌ{1φn kAvQ@٬pQ'B;41w`N/YVN}[n13v*tBV/`.Rԝt=jLv$NS~o-٫GZJa⚚W[5 ∺wWRԁ![զxt g}LHR,B\$EբKkA||-#P}ҮףL{{HZIDWq˞<E VjhdiAbQKl\eD~2#ᢡG?uKQ}A$%z=poR wy-*qɯ_U}qud| ԭ;mBߕ⼹ԇ8T*FAcn[#u_(zn9do[.+.ٟQy8*?\n%c]m9?ErUJ!9{LA!+eP7RTfRgH6CY9RԻ/ٜW_)X'Y`VhO4:20{(XrSwp4ZBKQښKb;1w2hDA5|$ 7:]jl"5ȭPSN"ؼe)f>l_/q\bm#<<}~zʯ\3]pT?S[6H6r k.k06ئ kMXYD8T():!X2&:R ڬGJ,,>$<"ZsO֮(}13޿ǥ5u?޽Ǐ1z4\' ۵oG';Ex~=7~_CM߄o-% :.W6]}6 _zw33K_Up?2]l~2_/?3?X߿~m/x}!ڇ=Õfbyf=amd8¡RPxL:%[lx c8-p6zΦI(L;GC>S, Cyfdr#sUyif dvcg6~$^Ry|ig>@`S~Q#r8Xa%x wa` }13)"/Gc7`&gՈڨ^M 2 O$LE>&=(&Ln &VZO%j5A4Rana F!-4X7;X/\~;j04 } R}*.~10EHf-jo6/j<̢gD *"EPRQ@MpR#Rw7;݁qnUo9Jڝ {(.q'ωّxnoN0 {Yn1rX˱O١x+is"'H0'8xXaۇo1MGk_>GWjb{'1b(I<@f%6ŬzZ@d w[n%M!"%uejJU!baB=Gj̬14ԩ<8 ) Bo/spNx&:O}mZt ei#&Ys]+lf/CX̌ ӠcS0iD$Mn#nj/m~ğ._mf>Q܏-B)(e)^ 178Ȳ ;hkZOO#6f=$)J}P= 8Ʋ!>0Nr '̏VCM@Ϸs-0C&?O} [Uͭ7XVpM!lHN:`Ck5An1q&fB!dH0BM"fZ(ߐ;1|hI|[ɣmmm M -/!"HT{J CҼ6wX3-g1V{ RbATQ^WRݛYMa eb3+ݻmeЌ"2Qo#b%+5;6՛9+a{탑j1}R`V'[v}q J.Wt$Z}K!ɧ5bMٻi3HF-Um_ sH`j-Cʄ6GCJV:\ε;5@PQ0R%yDyKY=͡^) )m5vjF-Gǁeƛg{yhQxڲ%s׮3r'@4hGCι^uL6`~o ϹuAuw_\,0 *p#r[ t;q$ԩEhDC-0Jxl)G h6 <}n6\=mku3#I>Of1.f\)[6AWgנdbz 6wXEkb_;s,H63ͮ<1&쌁Kw}|>i3Ze>=8 \,Z*e;vV sv%z2'i2߅·l_tB*#d,?{z}g̙ y]99TTmpO5hVco s9xVk bM>{saloFYK׶u?؈Cd3.(|*Z[+dB^H>׃ʎ%g3I\֩2z.dVBJYoX>(Y)W{D h\z &f˥oI>(zZkA!VQovr8cFQk1GPࡍ,qNA:$(iB*+YZAU⿑ +UĘtA1S@KhYϪ,0 °`Phd,`|Y{ 0?T$Z!ɊscnXj-+c"Fg3"_ĵ(K^;e dhY-:IoU=4Z= D#=M7q9!}̠ ujA칕}۸ٰ5-s%\F%-%okrsoi[Zv}A+vu5h <JOѿɮݰvC!&;=:ԫ[!I|NyMX'aC~vQ:GFPy˵Gշ dT4nMMEDMpp|/wn7/~ug)5)Fyu=oިMqQۭѷo|vqHh7=(kv rhΔDl[Zn͈*fsd+)ff Q}9-TM:~GHlX Hꀦk 5=cȎ'ĪS6:wUnA &Q;m16LDkvȕTe,f c˰ERWҾم9mS`Lƹ=1l%͆DIAe/bT< [Ơ.^!I;Gv2bnSe\f\gY:loSF[mLVOySb`&$ Y;DuQ(A1,#_K](lU45J&]]!^c⃅K)0:}/W!8ZŬ"E'o7Uf)DFDY,6IՕb6)Uĸ),r:l? O;V&bxм@5]f٣;ߝxpdo̢a5yzc-9t FDi߬drβ.lB92i1[h,+Y NVQ1Žl۸Sf# x`̌c]+=3Cbhg9*/56j,c Q$#!HmVu}b6dY{K1+u,8VͺmJZo"l^}2y޲]vxw);3v6dSsM<;!c:h>prI?yO#us!.ƁS+=dףUL-b!<Ւΰ]zwG(gm kOHwܾf$)zZk?3)B痌}_,mӚq^2=+%}6*oeL?6[w4K&ObK94k݁qȴr#,i34?$:Զ`C>ZIE*L^D^T:&#'vw?}a4&^C7w ~IrCTVU.CKw,J~ k_ֳ9 -5'm HoKz/!YÊؐ=|NY׻8 KPXMf\~~ ]W%9sb;kZDaVl(2xo>´^"ѯ8Z5>FU[EH!,agXc*dM\pkgBbdaS0 F$tWmU;oM\]|LE Zc1ɈϗjG 瞆Ŏ;g;QI8NLb}oW7+'_$7}Vjp-5rX\ G R8T8/HsQYf"M4Z~|zH?.^\/BH|棾zr"4O~M[Ik;kK:B%ƃ{n5itńZi./~/ZzNMJ*vu(]']ô56c[̭rbbS u;ӪU;_~aԿd9o^L_Xf /^gχWRX24v udp#w\+C?87N41uwŊ^ˋYMǟ뗧猲gWA&R=ߝSz=~^yǿˋ Uw M/ P^rupd ! :p; Sf DZe}QGCL_<*:-2@ؠF"K0M28'.r]ؑJeC%6#(|4 b2 դ[`>5U3~☺cĽ N/Q:<բDlͺxf,Ί%= ]`haշel~v.MAE->$X);bkLJkbXɑ g:h([Aܱ×fH0 h^.x <(.rh ڋ0";dVr:73vG%APOs84z i;;X#Y,飭 Ya5ZN)#T,jPY70 [,]SWϑJB=p\t! Fѽhu Mոۖ5*V+ J,By!6xDxFg}sU:V Wqykۣ[Dv*ajn;{&)t*}MY+Oj7oqBtujO+ѓuw\PF9mًXfY?QR*AG͹Xqb> Nx+\:6Tch =d5SjiXdB%1piqFCx홍evwG~W}$Icvl3jhu5@Za',f0\%se6,P$*|N16;c]:2"| ?(ʴ]2D ? RB[_d5h?-c@c/HJP~-sƞK9ƨէy*KP-.FL[Ji@GY"1}զ]}&O't"v+w$#2+&4jg)YB12)Kq>uo9\$ѳ6jUwHS`v@1ɚf xrk.BTf |7Ya6m,օ4 v=/A7m;%YbGts`ǩW^1JOQi@i/vҀ |ZߧO-ϩfijfjKp% &XS]ƄGR|6X-`-,Ɔ526}Sүd<HgrvMW%JEBj"+`?o9pd~ynzOuD^ 2CflD} ]1Y;iFɁ@ )!3).Ww;ruR+~-~L Fhnijy!1Ϲƨk#%FVEf;Bk8#DSz >U?* 1R}J?lCP-zn퐗UUo(X uM1*vkH2ONNN42& ~ry|u1 ֚=#{7f9&4/UNx'JjT_$4zқLvzbKDRa+cj}& M/j}'r$14ztXbj8$#6Pm!: []>KH]tύ,<ٜ4kЉvB^yj}/_> 4Tb`XՍ u+ $oK͡d@Rs;tmˍ$TaZ:, f[rA<ҥ_UΧI:LʽT!iZ5^;5cBV zaRv!&5 &rG1׊h)@bcHP%M*H ͚B"-!Y4w?EDXDλg3"Rc.5TpT⮮!v۰`'!$c$&m'>B*V jUlzו^oE.g1$G<r+ˊFɋLIxelR-mjx|=Eu#B"uئ;睏.u&AJ5m@WTx7\ XQ.h Dܛ"Jӣg*{bS3ߩf7ں]{SalLUUVWCU" d!niVEKN-3}t>7..e|:duqq#ɷ"wҞ-NBh `7JOѿ.mw/U뷍KdLy-绑4Ȑa[]>V#5\([ ³\x>}^5ĭ* δGx殢旣2,G$ݎy1)Ic 6b&Lhٔ;J+g7W78ǚmJ&e[rUgmy;d[.da7&Fx8ܟ٭.cBtGSe8.}E *95[V1|,q܍Se'U0nyL&[/[5N71rw y1ZP7bz^|бo\ZG#oD]%mPk<$TC^P|X u(BH+Y( \(4uQz_AÜa GR$abQJ b_:ϏgZKS絛:/ArVZT.!o qjN>dD̃e۫:eAͩ|۫RYJZT9f(т*Jfc?O!C.1T?rY(Z@PZYCKuDb a.9beJ7UJX8_6aއ楓Ǭ+h8}O@e&=n8gë̵v㺪Y9rՖiș6N+}ѪɄc:A ,s;x.gxtg uc!Do `>5]\O|a s|h*h"C |E(.7%U|Qힹj(vxÜ@jWTV#37Qwiv^Wl+JQvn>B_{),1ȨY{),naH{IL0yZĺb?E!)2÷X,gnzg(Y zFdsv&@s.Ǫㆢj'&7+}6`qC\Jzn 𝭔?r^L:$񇓓Ӄ?AHn]/]tp*NQ0*>;|:IM 5BgCX8B1AAVѠL͍1C*Am:KVq $q *zCT9MxE!,tz!1RR"ʉ ظ3?|g)Z;Uw'3a⧺3lÿY"Р',)f|lR1Kv,QYtTg@.(Uu0D9BL9%,7o]랹^>uAWl5y3[hvE}d*96ȧK%9DbH|i^TgiB[-7TKͼ ɈFbw.W-Uzyud$ht\;At\Q㘡DH1ΖZP7Ry F"C!cFy ΕPGď77~, Z~7^lY{]إj w(}bNSy}7Q&_f*j4:s@)=ְ{vGm̧H^+&}KeӮޟ?*Z*g'E?U͝ñ.h̢J'ЋݻNx{'wӦO^􅳓*ևbY3sdːkN4ʎk>t泓,^qz?טVg_cA&yi*`P*j֕7Ʌ|CQx'jQF&GJ^MU؞32biq{s@)8Z ,)7C;^W{o5_^d?s^ T}xsGhn~i hX([:M0!`m9Q+T|9෸Iu$zK_c rX}S{ӃKœ6үb(QiM">sI)⊟Wnܑ }1X(.0 , aYcQ0 sŷEg ؾۖ~aF1nږ~)0Y}ҮQWZX]v(C#W5E!,l a9Rq!PI}>KB~C_Vא?x=ZD~pztTx7l1Unqr돾V){׶#G[E',`P<ށ U沆݌[veV+/Ru@ʪ$#N`bZ_f )?.#ĺeQƪUW/SUN!{hzi'*$X=]>_\?ؚ''߮ęo~_^:36ID_"tr ӇA݃=E0xG1J,;:jxUTRյ%xPEb6PV]Cru9) ׳^,,O~;;?WE ̨»[NL.K:}ng#k1ݕ˷FQc_f`pGxv,_Foi}ؤZ"9IqN>M۱ˣy{?Ra't|N'n Ǝ8;L8<` s% 2QeBOϻȇ4LPWU]SS|IPҔȇ@@Iذ4^"WJrݒ"75\3Ch݊G@>ڀut z@]K~ѥ!A_|^ oy~ ]\ܺ"uPu-0Xf5[;JY=Xd\DI)p]HIA[{wvQzM=!z)jp&p BRHGuTeUȶ}e&qb'Ǿ2jdJє~ZOO|s~[*TEiW3=lN{y!}Q& XmA${|ow<;\s^\JvVw0 . )@]REM7VnR*nd_Lo8P:G)td"oݘύ!$!V3szBgIO%5s*R)bf;<3UX߀!$N8,|K`=$S1?1$1#dJMM[?} 84!hGWQQsͺ ￘rsɚyx84Cߋ$}'*ZWUQVoŎajZt(w,xh_4*Vq"2Y4 [\,c>#r]'L18Ǭ6be? ^'H|IG]F]ǻkf$<'z:=/r?O^ '{qm-` y]ݍ r#Ѷ}bRW)G/ S[;q$T%IM8>}12=Y">!B o=J-PC7biSٵJ|GBņQp4*ٯڞkj!?$1]3aagyGmvEO PjKXCEEjS*Ѩ#RǮ+۶t4MUU.57[=MUʜ*)^ Or/ȽKf](̣ΝД'r73ugaNw"%I`Ѯ#v;ʶhז+(\ ~8Q ΎZk 8ܣ˞;F>}&bH{Az@xu8j/)CJ49t(u1jgno:6F'J|J.B^E&DѨQ;pӠ/:Vvcx\v㱏Ƽqz ,Ig)ca!|PV`uA'SGf*6uth,bқr>g<ҘQL:}t;fdƩU"{! q%3j7o@fuB8BOϻJi/P)Z)9]U lRB>Bʆ%R%c]p Rvs*g6NU1Ysip:6O`p 7.$HoX~S85 kVw$6ּsV!a?u`p5k7n=Oj'SvAر)M: d ۀvOJ渃pӲw;.{VeBYf Ss>4:vy:j.}3f\<2Xv,tS|-p~8gM,f~kly3HHLq_,`ӱN9X@iY n'b ܶ in\@ ᛟo e@t. N@T1O`bt4D_T@t*n莘D'nI@ԯ UY㠿Ek4`k;,A'*$~$;A@G&FG[mȉp̋.hD&L>ԁV98"[wPX;0*R|'LEڍᛎ@gS xBpB0P;Q(T{gX] m!̡Z) `N6]ڝC19Ot2T|ve JR]-a+eSihP)bmw K/~" SdJ! )O?:FCagUTS]-{͓M=*%e5 )5A9AJe&TD 'l0d s16*s M͡*? vV Rx}`"2h&2)ڲm$_+|Y hKl&D;](w7#lKPv}ȃ6~06"OS(]}8oNeE\D~o"1S~7׬)61ˌä덕~Yr6o9O tW mSͷ.q0X;Lvd-8ÿaBu# ET*!;zI[:=V9{ΣG;Rz~: Pw$`g~Hb.p,n13-Mw!ۨ}gůig'N\[ԍEh([iqE,NىlakA(} A MT%Uħx9݈43טzo'1>U,,}bt1Mc#Gl (4ՄS!9٘CkA?UBNmo/*_Ƿ0zF$7l{Ͳdmn~~{G$x]ܬ"m `NB "P;>vP0)|;S]{6xx( r0m/\^oR뤀3W p%@ht@n!,JmU([ڗš\cjG22Q>:+x% vwr_M|~xwRT*&_?}zc~q߻pƛQWe.z8{^;h uEH*RSA_ u9c?7kAoȯ"G91Ck}(gܹk?j]oB~8vBM]UsCQ y{;~D딊 ]6΃.zl;~@SP\i;uPh\nCyqDzm!v{'2n:qiۘ&O#H/u8h.>-G[-Cu<ЫWSX\\˚ 3)I z18=iAO)Oi"mZSUrkZ\H!* UXy;;vKҺ@q .]9'н/kWnM}:83Z| gGtX%.ޝ#}oANB햤[]TRvTojtkVU{ݶݧۏO5 YCn3t7O[ڨgTXm*TsDIq|ڎ̡o5^8/CYI½RS1Wt7vq}Գz W›&9D{h/9~REq. T"F@No_ sq'{7C=d[ NIPǐ<}ak#O |s#%q$+$,>i1| {6I'M՗L>TRM9VJev|cx;eOif7E~hB:{<yz| P/^čˇ{Y`$;H&wLGܯarʝ:MU>׌,(3DS&9| z3jXTQ#v,ѷQ0*wEœWc"FMwΨsvƀ'ErNHD+pJw:*HbG̦nY\ul8Qۑgtb(tnnH#y<+u8/ {޶%/޴a;;M0Mikk$9qaV:(IQQ#AHdwWUw]Ź$!.GX„v^'c),5eNP*qB.)+u"1ƪQrLPlcɹ*<)-+202̖05´](I[cuIj(e#Ohr0)lIMXTYj2ń1`e8aQf>ENjQfM%Ow3KC^6NX(5^o|"Xpr- ?V܇XU%(2[|g`g] u/K X g UE5Tðkcٍ/XRtKMNY=cyZ(YCaXem!g=L 6ޟyPK!} :<4q6z;rIl𗻌V춁[ qk%w  DYƁGaҴxSAq pUS7( [ b!e+UHh=CoA}GPu]}2tbsxe6'`Pyw>]#(+W퀌sR)RW<2.vY knlu.[]Vo6IJW糥bp~Y#u7(B*A2P UY(uZ=Zim=Sx6TM )xg0n> oO$dލ_.rdɰ?REo?"@#eC۸p7H4-,;c9oK  An`e,k]kWC\h7C- 87Z4?4IoL4yZs> Bb!!F\E(I#B:~x_'8g;5.3#![l*=WC \=f5s5?p;7z \=Ä"&H^ 52\=ճq! I9 I[l E8f*vAo2OP^jdڜjgbdy,BŠ=++=PREyۙ隣I?;峫)~q ]r5a [,֨{8aưN883;Jò E'iףN)$K̪{ bEU:o/8ɽ?%6xY\ݽV-<L"> P2e>\3>{Ko0^HNq^^zr0,R亇Zr\^xE%X/;Oxg(77߮k՛1n)憓w[h5C`t qbS¸KS8V+͒cKzGeJd umc)AnKqxZPCb_gIo;v~'runiҙuI!@ ;0eu W߷,>4L)\cG}U?w"(%3:$: 0(n9zbq2ݶ cQ a( #hD)tJe#ZQKNRj.O Ȣ }p׃~}jhriB,R*(_CԲymvM"R;p4N߼G 4ם)jm]wof;BzxFvXs_~P*!(qG/jrսj~ls7Yɠwh/\* ulrIGwI NsM }_f@$悧r%O-uFW4xNwͶ>3t0Pi]842"iDi-VRs<]}\u˻^He^AYPȯD v"Nk" {l;BO[3PKK~FSP>%R|&yjTp6[J].6~ǿ]vW~Qs.\ͳ1]R#EKL$5/8 Sψ]@ϩǸzEjfTN |I1]BЗEY) 5aо~ 5%SM$\5(D+B[nED#%n>lmxggYOE [S0hwlO}j((M5C<F Zc#Jя##,4;nbg*Ol±upTȟ8M~"6@TV1AuR"=0D cE' ɈJfADb#̉$6窾ᅲcMK/li@ !b|+(盧AAM@p/ȤoNYpZV7y+CWQ'1^bBӌ94"#{[-{#jFyqb^?{jj\%r))zuJ*=i\| ѣT389+aLp<}er0mN-O|gr*Ҙ6G+A40|YzqKW'e^1.1UP U23F<_95YUAjF`Xk3',~AKvk \P vH^6~Y}X jķiu`~?xlV k a\xfW 3᫑43͸F2 Ӏ+WJ/J&§ cSX W6 Y=1͍!8:`VnRR:Er:xʋ0o{'M;qWiG|۩CCtп?] ĩW b"' i׽զCY@Mwjk4ښiόnN 0!j^ GˡbfhnGdžsm̑}|ó1gDcxn7ѳu7~\r@w@),wo]mZ_2aD Ig % ԉqtb'8ՅAG2i6 -ב<,g d4TkoÈk&zc} >ݬO,EH<oÈmbs*"3G88, Φl`o 3ɪ3&U%b K]4gc֎սʅ֒R[i Sql9TJplelV;& wplV{3B&;:2\^A_QstI`N}]N?ACz[aPyt Uba o ** ɗ1v*ϱEڳ[b(Ns&4?DZ-P`Z+c.+[vC$n%kl& MMۺ1Y>Th?C0lJ%NF2Ke1GE3|){2ZūOB9I7M2{w(wv}|֏3 e+e`4mϾ9y\Tc?4Qk+V !2*j/{S1R/|M+ *-ƑG$$IEXX%~WJTDzA9AWjM<Ӏ7@x_fksn6]ӉeY^mZబK(R5^<ӍF]u=WѻȬ>^:c)~B~D2C2^H?z0J_/EO3._H2OW1L{}|- ?Ys":"BtMxhieڵ2Z{5s.j;ALJ7=ەˏ?{}b#_h{@t>~.{^+ xJָ+@YIL,3N0^+%mfqiF"%SƸ D bE.JiLfqd@x7;E1ϧu[L\su<8ذ#Q}6`|LԖ4 Pp֦.!ۜ7ackࢎ'7 ^$<сgVxh2uz-]3wA$> J;Ok_~崐kOi/evA3M.^Hq&Ay'f#D5:t<,T{ѭXwM=xKl^Y!8r\q"?l".g d^cu@AaAv4N}w;"xVieRk(95ݴ>uѫ|c[n!*ΑOcWDi%Mc9Lĭls\Il'؊>_^\8wrKDGIuULSljVI f^iZD:a~j2]e0vpH(5nvmuתQ$"OE:rHT{AQc1sI][o#+_lN~1dw n}c;ؒdYlնL0c&*dW1X54jMBD8\Ƅ T2}٥i3(9f,:]YA*ST@ FlZl-Qн (叏- Jdi;^4uy,=RR E2q,?;O 9[mjAс*+lsK,}0YWU| Rծ #;<}ۮÄ\GJa. ?*,Qq)U| A鳾Ҍ~ﻓRBdz.!/VU)j+du6 kTذF9N$,!G}֑%w'*9LkdFY U`W K\Y J*B?qYNùg[&y+AhX])(>yҠE#EZINiSQvxwR/mUwB#k1nE(TSa~=w|"إ#ysOO97_]خo :~YA7)ʿl2hpHSe)d_舢b`vS^IB%u ,<Бɠgl3,vu7dRVVv2zK0Gr) ǁTӹTHjLeSfiNDr£V&!G 4a8F 1jš{m Q#!+e-ʂnYEs!\rF;Mw@pE)&3 1P<,Bpޒ(CM\lI:'V/ed.*;T.U |mC]cZpE\p@Q= )2>wa%ashJs/E.dyd{p,5&`$ڹzLLgs6G׫Fg|Y=V6:CL׸>;jh`uE=7c̐Sz^睶,BڿxšJHQ"j$Bs7Eu}H FޞS_{i3,7

3TXg8Z 6ӵfA98wƤA`]kj<2"PY緯<"TBjy@<Xi&1IVL85gyd2b\O?1׫XIjY>&]q?qyPcR߾iMGP0ֽؾRs8{$bF,%T}e ;(]ea^li͏h9Se^,JADCٜ4ы_OotCl}1-9EwH]C'ɺ{0=yG3i<3y7 ?й"BJ\}hϧE 0NS\&߹0o߹ZÜ15upo+\=|{k,BRfMѼ&wm> ަH!휠5Ob.qcz%%`dZ]7 meփ}`6$8tJ)C\k0@nf(^=8XT~Kp)YoSO>g)YTLi9*7SDOm,y1MT\(A!X#&b($ 2}޺OWYFzvZD0pm*+TUl=ehLT$^}4~xMIL–vuEy*0J\Qƅh'O7px $"NlYAVU,:[)]ER:i%QՆw *ñlβ7>*o=$5{Ҳ&X>KsfM31Ǘi% xbd8A~'e 7 \+(ŀ.9# EC%;g=~YrhB˙v AO  %%3>__O yXB?o2WɣU=i^4H5>vU) -}a1霞mr)g\ܜ1qg}ZDd#'A嗾o`44ل ݦ,W1Ό!427RHNjM!J C6<ncyR2L ơ3ۂy&β@D64ck00\,A$y꾨Kwy x">(&9чN5w_S318t1eTuP=[luH `\aSI9R5{{^|u,H4HN^Jd5n)rLՌk.F}$ѫ?*&ѫFƎ .Dz8BX((ͦW)Dz@kl! L\{;?}pnFSpTH*-[yޭTZpk?ø g:Wh{I8xk4ʇ{g?"q'qXjan0KV/̒ & ܉v*\Ze٦ [ O} $/pu; t ]Q8hFe+˰I%L Bq7+ERD"1/ PI\w֛u+$KnF's pas"t @J^W"ͧ`fGѷPps٫Ht'ȝPh1/__󄻮k/'Ҭg _ֻ>@Y0<9J ZP?1ݳЭ''pƽoA:}3.&G% z-~>'QK e Rʻ7)'JQlLS >&(e"kVW>T0:AW_oT6T]*Hv,\kl}P# $Wor^ąb&3YI 7>.IY̢[rrV"PMgX8=l%JKC)ueZûo3h&鲢@]6Js4LΨ"9h-Cy>pfV &:xkmB&0x6 ػ}~p!.]>)v88'(B).@kk8$ȡ*E$ێ:Q #Uj6f1~$t[^f\s͖!%hjguHcG;lf5՜v t3ҧ)vIA!!Пt<(1˥H~=JB'6 )y 5юwvFpzۅ9gg'NY4?@@fEW!# (ˍj -d 2cĄazLE]*Mݩ- gx)W>sȐtK,\GqưxgAӳ'dTN'u`pʦB];-h꓏W8SHPH U%k?=꿙-if?}p܍Os`Aܫp7O HhiḒ>JY}~%fKɧYgZa)xPܻ8F6Y@u <’)=)m6vI&%䵒IVyhIVDsWSNXD>ʄ(aC[g0QͤCt?ܚڳ F ^w]MCR{N@pG0bw* c INHc e̒" &n(QQd(ԑ- 1vD]gr?tiTuSwtk.4mq?E=(jAJr@D wԲyQ(1ŹxFmH' ɒ@"FqPY`U YH Lut!T!J]=hB%\7]s.G"DD X1a`85Z `Q$VF! ~6Nh2۽DlJ 8? wvfF"2w)9A'Jf˩xCg0ΦV"/DE]]񧵡py.vA=,< gk٢ ])<8F1wEb지h` ^>/nUY䢜Ǘ`Y8++Pv捌a׵͒N"ǡ.'`- M!c)nޏ|ao˒-٨g8=KғCP͗]IVo+ld*Gu {){cwTokE.G67ˈ-tlzY^nz&D@fbaF:12D(a6 Ƨ)Qf9|oάCwrNVA,. 'iozџZ"xg{/xُf7?{zvn}RJ_?g[P `Vۋ0Tܝ'yf{`~h4_E=(g=!,j٫bS3@z@SMӹ9sGgWK/ٷwv:쏲7%lM3u6xЏozן,#z],a!(_~ziLx]\~.FO?߈;ܙGoY﷌s KXgm< WWH`\5_Lǁ~Qƚ۟`WiO0h@yZt8}v)D8 wW|;HOo3U-F.8υכ,,1iɒ*_-e\6b/YZ;7 :0|ܘ0((uIflV(G^砳5hLOa>֘.6s'\t˰q+=N?b".}~ q9@?Yzvp}ޟܚ},|e7ғ7+#oOPWI>qt&ɴ߿B;/`ǀd W!kIH>׷p={[\~A_f| < uBIq>X_a Pr,`VJC\T.t!nd$w@I$[$"JbIG-PR%1q")P#@B JjCK-+ O)b?P 0_Rލ!E! Vr{T % !1:=W0;BO}jWG|UjWw(gee$0< F(@M:Z8DHVVU_j;_%~uPO.!IU;{tQl W"6Jni[,DXZ /[,bo(aS4ҟ&tX˘( lx@(VAbz&aK9[8_oyC13ZەH9mv%[,2jd%[,Y%|\-Pʆ*H"n<8D84 9%)JydD>R5+~؂Jt2-l1dU )ӫmjCBlUAHKALOK@ݘ"q!$jK:#v3YGaA9.Odݢ"gןD-lQd]?o;M`[$9 #( X̃$xl@RpFi $=w3YbIyFP}buP"vTb,_lgZ6BKop*epPEqzoi [rh,%R|X6Mlkggg-vw}ϾYdͣ_|Ke9~#q[߾o|_~X~{t: b'Pp`p$ib%36':8n-,82O5Wח]Q]FokC v8&\/`_ͳwFs[7/$~dhLG9-wsЗȏ뾗?/.^|} Fۼ|r N&|" K5)c`KyLh4LI:bmJRa4 BςJq$8k$r AsPpJLS9F0MVQqW_n/-@ke&:6\˘Sˁ,A$%±Q4X觹xk oӜ2ğŕ6!W{wAg,N!Ts=W8l -|S3 hh0:kH4;נ>6]>\mdDs{؂ yףN7 5Hۭfy6Kxj}9$#ɂyёY}j;p0 K`|jBdZlȋeVOT_]P*%Clb[(b1 ('XZ$ -WO^N-nyTFXVjoilq/]ċ8_ŗVj$@$k'eJiu'.7[a7 0+;NG ؟܄ԗ0z˵_MZƷoS6Ì7^O7ޤgw?ti8'%#_lWƢe rĔI,ߘZ$mb|Jr+@!JH (aJ6X!iƚ86Ք0[HA!9υHNwJQQ)JTyQZn1onYbUbŝK4N"I,.fTs8)LƔ$Lak˄lEQc[5^x«Gm:@R:\(֎ 5F`2a}$1gJ%M .M0Ce)a! ~ !aa= d=:dww-K^;!z !TtA\B&[;w_~){OipG[ #΁AL~Q+{@ky:oᮤ ns򶏗FΦlBa;c`8Bk;t6:Tв(]%Y lP 9.DJ!:v N*{, A6q AXɭ)Xqϵ7ko\[SXQ ,D*l1S%$6jN(,49ܧ)hZ2JwHȔa#c4vjVS63/s,QftzH2S_(^PDl-+1U*EDJ8U)։4IcPTqlE"%oRMMBϦ݀~1/--ŽJ ZKK2[BZ/-a}2,a{/h,:Z1޿Dw@29yGC Yې o y$ s:4y_0/r`)J'_&=%Ғ1R'Qκm3o"v*t͆7'aGDl88-6!Š ^ br)m9G/r,؁gώԡ'}|Lm18X;Uki]|R8ǣ@xϮL^Fr] UOtqn0Y"\iuD.x %ܰx8Tdu !}Yz?aTX5ŊwGW< ]̃|J(bzd<]E7Y ?o8p4Ɠ9L>TpL6G.zu4qQS]_|rEiXʼ.\@8OQ1ӫK{n-E!e Qa >yEy [gd7C9UuuUuuW}ig AƆ-_1n[P=J{h0њNwy͐6IeAA /zZ1$84G4Ii" Ogm|"" 1IOQY&wkFQ% :)Jjc{ͤɺ((@\`﯋ `{{ ߿x\.W u3EN)80IJ35TPPhrED7aEB Gup*I$<ҽr Ňx7k0$n(gzQV-elKf!E1gIpjΒT BN<8/^WsSK@+M+ϭj4SbC)=mHhQ"۫a]r2Lm;OVd5#2k!3І;c I *e*qiuu{?^JRkSx5ʏv9H b}s&t_K:ڔH&(9\ 0o0ʰ%D4 |c)6Q3FQ_:D 1ô.2.~"vO{[)]+~wߣNthD˷U30Qkpcؾ紹x|1+#ab^ o{{񐠢SN/%q . nIA#nTVQւf.`в1H4`c$!^Bmzx$ztEYJ8f*F)y}l8 ZA܇$Lh[trJ Dx* hRJk7C(g[J!R(.qZ1k,5RB(tCC0 hh@@톖'k e@E-PWDiʓAGO*QfboPbԜ! RjѨb9!N4)gD NOQIeH2,2ۿ7#q^H Ad#V{{KR+{ƸpRFV89Z/s*QUsK9B BͯZ{felgM>+xeo|=<^br~}.~yzt.O8 Z9 c{8%ڨc;MRy|r/='k)"$T6ebDuᇕ`|ͬP_@v2HthN0! ,1U0fi'h+@9SH}W4mdL}&@$ 慨wB?-Q;^oՅOVѸw:B%qetd8^Ý5j>Jhح$Q[',H}jo -HOxiJX)qr6e:Z:t:~(e @Qv(}~6Rc*Ϋf''\OSj7ٚA!10Eo7/=0+p`bkR*5F$޿(d {Xtuk+&l@(Dfm-}w)JDds3`CnhVdvrl>UNu(Qd<^.yxN &Y;3wAE͈UBVGiׅ05P!_/TQ~* BYٝv<;P1,|iZ ]KX j.P#<~7gUūF1fwɡ/j<M~yW_ŏW/y ^ 6{@x?LoVXÚi_yuFz6ʍړIeFWn>X=uEd:bָ1`4WŮ7y"/v]=_iP[2QFt٨W F+52ue2*ssؤwVWH!gvZ[A{s×Q&~;kn]AKS;F{7TѼ[$^wEB4uO}nN]tv`jBK[EJ􁁗 5Ϛ??3a#]>Ac\ 9aC0y/IIKQ>_ 0)"|1G04U1k>ٵ 4cUSz<ctxӄq:MPyABV4NG'nN]۔,'d,jBK[EJ聶48 "FZЄU4&^8H+. :XEE9P'#M8 1*"" p4 '9J8*{~{g߾~,gNov' Q_ޜ U Pr4n#gZdy{yt[{չ Y1xrׂCԝ W?*{uuk2"Z![:޺O1[ Sx#0LCFH| K$'X'ȷu%WddZXlrcIzӈ,9^F;d㋬P9RS~/Tb_=cAآkD6Xswiy%*CpKje48:$%+h^U+ha =Ȕ}Ol=s_UXJh,)*:JDEٰ TGG&%f%},mc?Y"HP5T]#0>.QmuLB KV}?VPdJՏ%MŔ!>HP"8Mp-\Ĕ53 DFsjj&멙V'Lhf-_AdUGJ<Am؋kxm8i`Ym./W&6Wh٬/\.alQÅ5|zՕx[C^!|՟xm4}@5VmM7fpnj@\]TW8ï|{9H#8xiq'6I2ۿLChm eJ*#d x,-W}yI"F an4]m)bQK4wxmg խl͢+4 ă;!m7yKkO@~@B_ȡ|P=~QQ4`U/D+ӸUYzwGkk4cM{(8#CS_ջF_D! !(1LP;L]OBm`Rs KI.blڱj˒V_N |i əgƅ`tD唉FF\3=8H:Yj΢o.w3EIMK{:>;uۡ3Kpx0TL|Yfcė}_2%oxNx/K7B|Yxr5HLfT#]Ri~Y|wWl,~"~@{x7robms$:; ~\B,?ޚ 'd8&I6N&~uzd1 hр{$ #?1XX|c;ߦYyܕA/wbkâ$]X/;_~*k^Y:pYAS2:_Wiw|69ZZ60A! A"eL6QVd>--SP$H}DhiY噺 ]cj͟pu!.W?PTs 7 %U{a% ~Xd|2 n k$W!JӉT{Lwl#MK\"5=n[դB늕ѬFgFo4A ґx.,JI rE0bȢkQP-q703\fG~ϚwqHmUI|MAn6V5)%e:oM+չgG@ҡ.)0ܹ8w4J!>1w~ֲZ %&/,s,V}`ϐ -(&V\6#g!y`z̡֕c&^yꬉ/B1!sX&YyC3*Ξh-R F0暟Bq(Z94S "4#,-&M˛7y\YCr#ҳhL"spk'ȂpZZ"FjZ|⋪-U|,S&To6YN#2,Dsn!Apu4|9oohQj%Ft@h(V%.Cƃעk62 S[+ y=D`9?oR7oxh]x t I61v qF0}>_b $% l(;;_dBnf 8Q32F G4yGP.(cX fF U &b|ДFD*Dg)D/ 2|)JIV$֓0łԞa, QI_ciTX8 avnRz3jGr@d$Qr "Q rI, *\2mӶ3fZY(e? f-aӨ#NhkeCo\k#8Lv0d@3vMY.# &vF7i|>ݤ#/]P?차y?Ϻj[.o~Q^T>^| ڏTOG/-]C !W#Jؔd`v0|;i@Z~swG"qHέL[X'{ C8B$8!g!<*I9lwu8<-|v3Xqq1㇜>aK,06?j.4k5/XZp <:t ZAFz鼅:8"D* *0wHst;ƹh[n<XnU`k,2yCbZ x1*/*'|316x)hwݻwЏdQ{%y!$ LX(WiͤHX'IpqgH14bAɥᘏ0p[wEa #iL= vܳ}pb%s7OgAq%kᛍGĕ z!W=#ʅfQYC IeQ1y}Tiq9NJ4 LzZ*uF9tJ1o7Zr5~^շW][͏ S\ SA-DIzNDgR'A$Jρ*\‰XG-IFmM?9evή,bJ\NQJK _ i!K@amnþgn #N^| Hr< Ɂ7-WlJ$$h&pzf^eHK;_v&C2tf#\; g1sV*Qq%$r+Si?n3R񧎉0LnCH~}j\*LѤu6Ap҃6Lξ` Wzdd\\mO.^;E3uM[HOa8{ZN: nTCf6Iyj6SV#/ajJaF %!v:[Cf~']-3?/e??a }4`6&%e21EC/_Sr @/؟=`ѝ3o9 YA2§:ZgpW@}53]9Ne&QWnzk g.:GC^Ԁ-#(l4`Tz0JYipJ"a<tb?e>Aޯ{FNߠTTJ;SrZᴿ ].e*>&6qIa%oq"I!另ǼJ[i2l񝹽:gT7tä~VSat?N.# ~st*e4Mhc-p#`{8QFiE",#w0udRMnCm|9ƻH=l8kl>v0Dpؓ=r$:c { vM}^ro2n={r!Fd̴]mhY>xB Y?huɐQ2ڙRt0Thg=G5uJY;q!ZEDպ..ܺHЍкR4@[3 8Aj&Ң:td\P L^D}l+ bK^{EZq'qʑTGd);JTJ}*4偃o!K26xbg|W\MvAz9}/; MWCz UX`Tϡ%۫׹Y (~Bs5)į^ܭ&9;+=O)R5?Z>n}Q"H20!:qkLrƈHd*,V0SPBs-zˑ/(]S.y8RDR~`hjo'"Drye)GBԊ.Et֜Qw#SzUACf8 : 1#ъ\CKJ?Cf\ ֲQ m.K]e_}N2U ԝiPYzk/;Ufja_fq! !! _;x\ j$ TtT̨h^9lIRsAC>96-'Znv\yOːDqN떛q˥anlr:@ٯfX*xgA:|8@\ŖZ%%h "CT?gӍ5 OjpAPlцgRs)dLsbd8[ "Lsq@u'tLQDQ IL3# ю޺eݾMӋn-XE_w;|3_w*˄hAS}s!^yYìk ỹH2jwǸB{3pzp ~Mx7<`@5Gͩt6W뼾 37s;;QLؑYl9:a3笚}'.߀{ĠV$+Ϥ½0/Cy( ",fѿ .Tbo-b@5i+rȣ!hKŨz& d,˥ZQRGwGli8ȃ"` 5磈͍9P<8dg=t>CN/aLIWp 3-!HZ\:ɖLVE͈ '2Ğ{DP6P Gb D:?+(>gw 1;& [!Z QFBYAt4!x3@ҜD %c-^>\r-),eZ6_伽qJ)2$HIXԥOImS,u"0.u5UFfS&X#[I*xTKX v;ٹΊ"4~ln^eS&eV$̓tw\ep?^c49R"H+"RRR炭VSw"Rq+R|m".;lkt3z6^u$1iM'Wa\V2oZhL'(AnrVH[7_)޼sqsŪwhRT֝Fs`A2|.Si}-_{ﭛ)4@9P8m/[r;'8EgdNiKw͐,_y+"1iL"IݥfgXo?{6jRkx?TۚdwRs3$íR k(ɉw*5H٢e9"MRbPqE}@h^ |6lT@Jy+H r}B @M+Hb*dz!xJFE1ldw XJ ySih!: Q̝ꝛ`KqzZT kMHb$菦hMPIhH`~2M!sm$({ҏC f.\$[m^\"%XaHeHx:Vl$ Za0c6IM8fS& fI̦Z 2GFY#˧kbG^y\7<Ƅ3qtm`(qѝ0WfRL2y!h} a!zUB?fHLI%W*qP*8G0}D[dF($9K$KqF:LX ӾE%a?)- F$ EkM= vC*V}@l \ì?YWn~3577~`ҕˉs @e6ٿnbF F( Vσ!?(Dwol]'.)j6*>޸enfߏg<9G5ܟ[eМ_\Gۯ֔m&?Yt\8; _KQ*}> 0KxT\€.^.Sfȯ Sk5vf._w}1 QdW Rkb(FapJpHM850g&b$2bxU \cx|N G3RWr9 4I?QՊv>]ghc dC8[ގؕ.F]i;NѐKBԩBr"I!V^?bzϽm8qM tL5SUWjՐdC:8!qJ D0pW6!z:tHT衐來@UE3Ex2vj`2o B1=7fyl ;^\%1lfuKpR"'id&aSQڻm?(w6ĺHɴKAHDJŖH5q`? fT`Ŕ VD:$N>+Z:osQ5FCk!:p(l7:j!T &k=H>ؾzH[y;vVư~~2Zb]:s{kܠIDǮuzL 5b;7գغ\(R5?Zdܜ"©'p "!6vĤ"!d l;g^n콯(8) =|7%_W]L9Mʼn,5PLr*[gY`Lwk/ZSe"*S"8jE>MJ8ʷ?4"G ˑqt6s'q^sP:m:dBj\Ehjgh9'A{;DJRC0/qI5)ϐ L+Q>>sa)3zA "@el"^YW-(ұ,_'c;v˖UݗA?8}ST9`b>WjPEyq0NM %DYNƒ̂B&u``vo̖5@833}';JgVt)]iljwDI:Μ^ÎjYP,/ZzL@t+P~xWmF{֙lX\rp~zQQ} W f9Pd uq#P}b&k͸tf?ho=JmGrKiz;F,9Ø}XO)e-[%不@x:`;#[t U"zs'z_DvXL "{C iQ2O]5=t;NoD][fKa{?.EzCx"ڽ޵:DAh_wYNP^@ iMyna4R_vUuow%=C\It;rN߭ɦ}?{€//*td / 3o2Tb}(λ}9Z *Zu-['J/I:]P`qX ?A: ChO뇧 *݉|˖nӝM ٜ%Z7GF)9')8v35SPC8&n-ɜmh qSܰIʘŚ)I ZHCaIdG+ȎVI|t2s,Qot~h;_ZXHs0HD0IQB&Ď3!"Lh 9 L(i ;Hs4'% ((ѭŮH|=2Z(cQKU0;5YE:ؐjC T$tXOߪDI?'0%E65 ,nZhM>A%UfsGA_jڎd`{yT=cy!;x2'nN,L].n.Y]=?6|Ww{Fp7}*0j (R̝@q ズG@CA0CM؉Kcj9e*TklybL$KG K62&.c%SōU{wMx(ִ1IP^'0qԚ,;kQ KÐ3ro.yF5?l=rfyQJ{j4`X.M8vuI L:r`4AӼkӼ!LkV:^C-ML A;48kYlxL|J:LwtK4!0`|q"@} 4XH9QF0M% dj&Tlj|ٿrϰ8!â1`'18d _8K- $ SxP CΤd<2.NtjH!ӔXmӄ:tC$ kÝ סø^sJ0yX'ƭ:\ahm1G4G:n]ڦu_SA N~;(*g- N;( '@`C3"ͨ$_A;&FnHXD O^L9jm_X"f,oN>8PJBu `"Dk,XĂGO4Xe'*ـޑ0+{uKg_Q-M)K!hc2M,nG~f??x׋C~-oI_Mr+&vfv<>NXGDE[Wǰ7Bҟ?53+B!P~[eή4/nXekݠͯft~8ʶW< ϟ/rZ>m4x%h/O|zd®!kOi{l5g[R*E h=hh\CKw3O׻r\ F/ '\,$[]QQ>4?]BZl^V>͡cDQ ")ES H;0T@e9MQ)K+s".-%&<"fV$ΩIu| MS96KM%RBOT 1bݔ0&af;*+V_)}eIR$K%B0FИ@LI=J29PI{OCho3l,w FrW;¿}Ok%ǫ|Ȁ3x r;܇ߋFB(>~˦!%pd-RÌG+ CSˁ\.\6{wlA-Jr;LPDL X[B KELӔaa")볢>k~krLK9dH6Ă(N &ҤXB`Hdz;gN\ʻ+ گ ׃um=;gcy4xs+ x+ ph Pw /4hBEJl7pGD4&ȖObq:odYO`-5ouY[-MfgoƿWFp=7!gix'(F.2F/}YiΫŨ@.A^e%|Ys`(7E5/5!rV`C bu% 3غ6طNdg[RWzݱ_ݽ1w~{+$(u A!KfJ )n-O[3&Dt5 :#fO&sVdq>ŋ&>d(G{w25VRɱ iC#7߯W+8mmU~Zo JC/O]q %IT=+pMѐ+ E"Bڇ?j.S0*_~x1f#myvNt_eu.C-cϏUd(R{\ @BB'ot?ynbA %{ LY-<޷U)wKe\ Cn ZakjuL _m+yxKŹ_:Ab w̹YR#|b,T:KtR,`HT|M+rjV6\d #e^w^%6y^8sW9oooo30%O -04[Gt@OL04?'XV{iԲQˑΐp_{-qIְ Dc9wZE wvO2) YjoO$?oU:} )!d,_fO-)];0K+ǶnEƼ hzP@`:s=h6H8P9F{T4JSmإV=5wnbucPer4v }e6=BR5ϲIA͒Dfґ]f8̆qnWk!hUx1Bd9ՔC$}{Wv @G ,0CeiD"_%6P-^oQ.KB#A7ynn7эQy= ( 3)igƢɉ9ΈZY, `nU!^WcC{L)TϚO(^^-'3j@o)RMo6p,% nt纊!(ѡn2SuTc_?5`=:D*jv=\L='$8雷*@zOgt{C?W9G3h\Ƨ- dotD5le ¤NWpOIhÕ@pA_nBG6YEnW&bm\ t<_GL+`Y,!U6tR*pN C.&~J7 I2~ h6w꣢ .9beKi .,N.$`PGau/*Q&5Jvِ &{S ZVMm,!'geįl8KG>. \27Ea} a# =Cܜd4NVG@0b}Վ\m d3V|TY `u)z:hͬS|HGkDj9!0^[ן2"1@."-#GZ!cX-$a`x)vQ0 n-2XH*\+fr{ '%Vpl:ܖW΂XcNslsI;݂a=Xbu:׽7b!fL? qBHAg.{!XyK q]EYIViѪx%u)e#=N 2g=Z,3;D"%`faHyϢ2'jm`8'P="xi!xt(9{Q$%qh[^KPIpJu&G9~)6'W`Z\w\sOq射(}P-uPkr+?3-QjJ%JB7 +$Jt(){SeU9I8Gmmu 1m 6-2eͨzMx}m=%ӥ| )!k+Pu6eZ^AJCjwFy{A/f>oF4=t[$ȷ_+9 GDpS?5j͆BX cCt<1C}pA4co>M|y 5V$oUtU4,DSQmۨoyM+6=S]4Mi:1 ߿*OϲxN)`Qb Q1ԏB燣 J9f 3f[f!K]X=|6Q4^0cM;^\vnL`PP&g]-rl߄rR~{x\xus  _7ɛkQ[6rVÁz8US]\] `(,\? hmNcbTt4tV(p$}J-1`=c`r:;D@5A`})@aKAŔO%0\^jg-e/O8孩00Վi'7giq%Il#AL (.϶jEW:ĥ7Z(Ckء"UGS$3#G S 'rn@XU,bB6xHݟY)h!=RGe+=ČBq8FBo&=~Fh_?n}YLj,7Ӡv6ߊk $ B z$z > "{RL{\LnS?ra FeX)e)CrcןvqzDz `xqu47ѭs)0-ةy֡;|{en: ~4^wN_l&jYU}{eN{wR]tym\ĝWWB|le1gf 'P=Sm:!Xz|Yym+#g>#~PT/55f$Yy au=#]U")aI]SN$N ypw ;+eQ^cLt 1#j:gn]ZY8"2ڮ'o]kfdid+G5rɎv-f3[wɌbb =(k"! 1NV;ce q,&qMJ\] "s #!F0@}8`=y 8'G樻=xܴc>  t:)ZdBre.5ޭ $Y,6}Si#SArn=X`[^F.؜Zϓ<8sSF@= !Ɂz.rSH `ji<%~y GC _&6ÇBt^W`(g$ZD8h7z׺#NaoSm7IG2:2ci{3+KKA4Ah2esL& &InNXOi .RBN ]Uzr1V!<(X Es1m"޿-=TO@uco;F N ejb||un6,xڰsx$zLe0ZUz?:vG?/J(J\Qw2,=E*yxRlʷS҅MZ c{ qz\RhRcgwz^@(y:m%H"߽)1S PRS2rYk"9fc0FJܳe=D RܡPXc{/ˢUeVB"zV>=s{$7y:7j1dix8z+"g ڿ>@ft$)6 ,J$R /b~]ax.ps/nxt[VO6#dכz :OGnv w&@4(m;l3ΐc ճ/ s7'xip&qRUۜ+z* AdSPʨi6`Ѝd'>>S 9~k,"h?İ~Ȍ[́dRRM-#mvYT~m=*u^M*$YǾ2V|.-J7q uaLHKȜ=SacB%M|a,m)f?,`Pi+ ?l؉cxh8]+bGs' <0 1VAnow*0Hu}c%-_i+.EB&&8UwNߎ}"; Zҡ/o79iOpAZv)qZ Ц*ӡvX<$@\RE)|5%/,Exu*Ԅ ҕuP||n!};ȢWoрٱ]iV.:4,\;aT,e9&=hc}80Oi ;(}](n2HKDªL]F2peC[D%eytJз4l6>FlJ/[JxCI =DQm~Hqh[)KuYMc8 _ ̰ |d4դa=v`369T'gG :|܃T-Q9' .4XLP>2@0W:vPwr\?t6f~ԩb5%GL2Z !w +7r1˛,P׉eD8( @ʣFN  tA'Pރ ) o|◚'r*H$DEVd8>[;H S{hvDXU&?%9 =1㨣3 IF+ףd6kRbL .P#6Z Fؚ)>{*/~D~<}'"{t;c+!? ˭Nj3W_E j'%m(oUp -t}^u-.Q0̍S9kΘ}$'Z2oZ_7=%%; q%=t/tE=.}ÿFkc89x~o# 2\c|dծax.& 'n>"ϊ6? gJ.yuscr;eGJo,s"e^˻r#~*)bz[jzw>拝n cи[*bJVdAܐSOXl +wy!8ݺ^WM"ԜayPYs8$%m8;UY۝kj0Rt+ NC K3omxV*Dd>-KMz2YOq;">/$gWQ(/2WGzI]ګ(؎BfQ8y%LW1ptڝ_ct֔˽s &1~_^ 'T9F*X= ڋՕ_!gl_%R3ʁijƋU{u䡬z0IOV؋[zzB*Fi>?i>8"{1+8A(5r>{9f[gm3͊@'o}uT>[^²6a y=",Ank{YLzqZ@yxFE <0^"H̊9g\c%"y"N+xl" spd8*e()mO\.uiqg *t[Q-j:J4He:;χnv󸶛ǵy,P$Tdꔲ!ƴC(ӜR ))8%H44@ Eؼ-xr>.Iqw|%j:YdU,f$2bj uu1YsF[Li8YoI =;zC}2 K3CLRQC;B3Dsk 4S)( (,[m kؽz8a5L`&)1<gZC섦 b "CM 99 BY5@HC XzrJKAT>.lPtYu0tf;nm@HLFNa p@xE oz`lPc/b426uNR%SkCR-S G$Co7жIm E=8LMBO0a  VSsA\$OE nt1_"M=nBבfآA=c zt\oxrvz.*p?`س7'w\懪|lu++ R$;bEdПדL7"y"G  "ԣ#3ًՄ;m6O'X~RŞ-fovZ.%My8CNf $Z[`sís8R;:j7`Ƣqk;Vn_g )~UI폃w~?PAtEv Fԣ4uHZApr5{Hrll)K3T#1j(BU S A $7a i&!wD즁g6bGasGZ.Zgȴ /xm3:yӸ:) Pbo*ypwAϘN_kmEټ_Az&mdv/ ^ۚ%$w'ts^*VBts!yo?ǻo|||4/xDЛ?a1в"Ic$zGk8q{#hB%x2e2_f̿+_ZCkmQiѼ}t݆~d+ya&_t4͂!{ 磙?~&>-lh%DܡL{Nը%U%LԄձuuíg0NѹAVF \\Zj*51$tw.=2#⺡M)J*4p&n2p NSn h  AqN-5v98B#r}l#|֪13 3V`?EeprEXutdG]V=k$%Rv|Zg>'>|80YQH1HkX-3r0u +\>jGU|gv,B-3v.WSG#)H%|I)aL6xI A0_xr-tR ^)Ugk5\%:gKHּ՛jIk+ZupbS6^9X#6 W!`XTr҂HVΊ%\jr1 l`yf$2c0Յ3>=,;S?*]t}1ӻǑ[u}96DV}M("<Q/SB ,P愎0y40=vX '8j0'odzYr;-dnK3.ƳVE\iLO7?9Dӻ;L-;%{q; <ibQ0f ?`egY嘅NN=SDžxq*tpf%߬RG ;EI|r,L, c;ҽ aLi< Z.vPnT8֠)!`ՃpX`:h)ȣ|EQ3ԅY$FǬBTQ q(Cq7aN }!QKkYr(¨)t٨̵Al<"lTkt+@c|bUp¨5.Zr"**l,4.xvJLl5L%p%e* l:BbD*Q_Z:ɦ7zwZ3)$>V5PQjԑxc(,GJ`g`/5T8(8/7ef9!Ӭ7Ɍ&/K/?ˀ^b7&z R gB r(%P&kQ?LpW٣VS˵8xR@ ~=>f6\mVRmgB'V!Z V mIԲ?J!P5;\4 y0_uTQ&_{8108OBrD e X`A'e b4鍅`vL`̫х0gbJAB4Zƌ d1lUJ@;8mhY0PX1*/ܘZRD( P%reG1 [FY+'+EoȨQZŠ>2(YK Qd dj c\0ˁs >>Dxrwq4>}}l:Zɱj?q~c,"r40EAQLjZg, ro̬G! `YiT2 JpQ`Hgԫqfq^wc[id}sfG 1EQg `5K%a5EC $ VsW 4g.9D1B-mdxJMm2{Ir=^։3%qLsÈ`|vCI 2cB$hs:R{|iu#}߸'5IL4[BpA0P D[ PgۻE\TtBN^>|; \DP0"{E*5Z+Ŝa9qc\<8 B:hǍR:%mLЀ@- #{qUqu_V]PY[xPm E8@I(YUbBb9J|T'DM&b֒|Ӝwڊj!hn3PBP;;i8 j,0֖ty}$Ơ\<J Nu4};Y S:mv*GWI.u۞dciy_ >\-YXvYZz/$zmPEW] 1RH+"r){l3]:)[~5eN><_>i`s- bDE#[4Ȓ_t8 'f (BvNV@Nmgb:?`[}V?[#}tCuCp6qվpffL>ufkvOmoԆ 5ցQ6 ksT1+бQ;?w0 ?! _)-v.2Ve]S/QsM3w;fA؁8(s6NA:v-=оbP:*#Gn8ݟbn#}1̻xfm5*]Ԩ9Q;s{K!ȝ>UNZOatmoYߊus暶Fbn m):sI?xK7)ύ{>U/Q$a< Fb!CA 9?zb?2wPˋcFS8em\x)G.W]%'r٭B}Pf䜺yFvH.⼳B;J^tӼ$T|%kAq٨տ8e?2i](3: /Ϋ45sBt~l4i/+R<+r򉤅O$B&Q]ű]*9aNjMT-NWDkd 0%s-ulsuqa%BDu賺NSdK;=2 /!|XzS *ZuB`4:W0 &V_6Kg'\M4\:)GS娓!y}` Uݎ4ZS3%VgL H]JrIdEɤ(Zm8aRM,# k %6T/) գ?]D/%Z&O(\jTNˁmƩ̖%$PDIur;J֍yxܠR3 I,>ev2t}P>NZ*>nJ:9!tnG6[~69!m'qS^D',}~Q Ηb4{<?{t\< 9ŀ?3y }}s˚a}b3ү5Jnm4tQ4 k\DJ7H/uSRAYj:CO:5S[ᾈ PTSP T*ԛ3ˤ/ttdTdwbJ BSBq%7k9u~ç< 4'<=Klzw4g qW(ݗ<{W3g ׄw h;9 ;Fԑ.I3dD=JeS"QC\)PrdsjXfOf)=!%UdVc I]dR+3N3Ta_zq |Ϳߣ*\i`;5`\=|(#DYA X/g6o `p"1DjL O0 9AH.E' `-,HGKBg LA"DY'!T*CT6xk@F0c ^ 8FL0p_4 "[S9Ck Wa`Jh4 &1LD):l,RIZ) L >t[(j,DBp`>0D[ xЂaW*-f@8(iP.`t$J 876uGN^UEliC^ _8@77!|{ 5FYU›fnǕ]T"0mo D8Kh+I+qUm2w.Z5r&:Uxe~_eϲ~0 ~ٹV3f4y9  忞lp>Ooiqc~ u/Ј2lqihPhaQ!1 ?zYC`PΊ粫] kQvmU&9XKY|XkQS"cZIۜZ8ޮu"MTyãmZdgmw/d&‚@ q`}O,:GsѸXQK$l g`Fp({uz-iwO"fy{k<=Fּyś1~~1wk~1ͥн5rfb+Sgʚ6_A)J/=N*uy9JXH<@RJ@%D%t PG;R_Ś_Ű$_3ݯ_(>YQq~1cdbŌ{`<8i.`me X-5`!AmV\'O8]Ir~1tb}bɓ9&'s,O/s$dk~h{p Ja{\8?5#5.`QĬb\w/婜LSGǿw3kCMQ2sdn(\-jG6 UToBPjelXΦ7i$#^} J Okяf ;=9IܙW}8-MH_΁X@R3O)w&Wy.@$$p֞/S늞7_d1O7("xorwJʜQCƨ042Z 6$F4wjGtl YdͩZ`|%/a (S@{bp`2Eq UL{>,]c}٫LoOoBcBF P.Fx`Ey-YUf5YCcö"P !薀JNmO54JFT''¶ V;j#SX)T+^3˫Wzt<+ɠe۞VЊrӛ_KmTkjqFT'㶦7JԼDTKNmO5SlEIM^mO5ZOfUji{"LRVx-JHg]VF ᅗp ?:]d?S2~eqƚs~`*{r7Bs^ƇV :o i/M{5˚(sh>dͺzK tRĺ/6ՒuKSih@ȉC4 ST|cݺZ[*@'uMR SIά[z'ZCBNIڶP\u#ʃ&Ic붿u ;n齦 [9q`"zu&ARy:clbW&}-Yn֭ 8D0֬,[*@'uM붊&ZnEZ6rM#{֭[Gʃ&IcE]4}GXCBN bZnvѺ tRĺ dv绥lh@ȉC4 S\|ou0:ByШܞ8F!2]M;BZ@ȉC4 SR|uCnV5Am{B>{z:[YHw=]Og/Rv ]/H>p}bu}]@ Cwvt5ܪ&pƬuwvZWcnY1kDvm՘O rt5fa]1 1k~Wcj̭jvk DsWcnSp8w5ܪ&Э'Ƭ1w565mwYƬ9)]j̀#8o`jþ| 5f%GWcւnsWcnU!jZ#sWcnS4Wc ՘[՘;j]=Mj@5aQWc>3`Jn՘!\FZu5ܦ&b T3%sWcnU>}@j]]M?GWc#|5f`GWc&bsWcnS(Vc]1 [ 1"]1 S+O+ND&ٸV?ō‡"&yLA> F4&g^ݦ6p/>Lϛ<|1ڱg ݷ}=O޾("εElhz;eyԏ=+~GȾ( ڛ?nH$ }+AH_[炗B;AYGQz\DM>.)2A{;E ͅ a>hW%sz#bq,DV21 #0KPAb6pE%4\[s%2*BJx{j<kWj%8̛ȑA,us,$:0fSS yȉr9)+I#-SQK`ZRQAF*g i1ZkdX++PgMdamBFsfi܃rV<(M{JV(L0`iʜBB3BX&, ʽUA@h"XCKs(,x! "1pN9 Dz,)ޙ< 0RIJ Y'¦PՒuYb0-{+q0鴷o.z[04pp}sFkg~/}-]ȥB&7CSvS~U^#z7 .¿ݙGXF^̅Hc:\]ޫ0 ?z?4wMD!/U/ bZ.pE>oo;C3-3 Ϳ3^0hGJ1*]K|aD/o?OD. \#kCriTTT3Mz7+Ho`T,NLt00w6Y=}W62`&G  i@uE1:+|6Yyu,'/l2GŪZ3FqܤgZHm c (-V$X-V%ZrВ3u&NI}L[ܩSy:7AEK;r2 »[t]߶&W"Ÿ |[ L>`Y\4o94HL$kQS?e_eg%%64Ɲt1Sc,/(ri<48r-%*%P33t0Y9F&YljF.7b+e?Yu G~;o=4 _x\z۬aZst0]}vP?`*U^HT/}c-uw͔G3erq\s0-< όu#8Vx/$sD6G,ʁݤwk=d֊کbm DJX[fkxySW̢A]!pzޛreMg.1%|'J_|7v.GXpJ5k_e?Fcׇ>ggf>?ᇁ gpg7I 䞾{R|?Wڕ(eп,W۹W ؽFR?_b5r4XȎ[p`U,, z?.]1K̜A+j/(ߖ ;wX6pNsJgJrdgc(ڬ:4ꗒSXWdfB S8O2T@‡X[a-G9RK/єҜa#GR%BBRm4] ⓒ"Q<{Lb$ qנ3 j,9fS-7b$P_J3j=6 )`Zaas.nq83hK)Lf잦H8p\{}y Lꪓ0SEj> ]8T߲W^7LkFEK=f"nm{ڲe˖d;"GݚIalw^٣!>X- M%\A.\[g?&\->h>WL[(9+{[LJ@pD9ʃu/<߮o*-=k޿(!$@iF4V뫈hmƸ+Tvd6z a/q:[V^gSWk&3< ݗN[apA_}RV<#΂3=Ϋ)EAciTp ,}w`݇S׭uM )Y Z7U.io-DVF}R#nE+.:&$GZC\SrD`&^AeߓER@ha *2ζ@C[g'QEe$ia5F*[nuhmԮz@}mkG:|x>Sv.cVbSls7*er^Ͽz 6-a޿omg:h6$PtJ'Js CH`> %˶!bLp:X4FGq='z=Q=! +$r\-N))a|4DYK(UKRr.ml`"1=U@bb+,q1@yJm~ao#}\5UvIQ !dD#Gn5IoJ5\tT:= ?9E~8v܁rħ;*6ЏRMvYQo݋ n>=91䅈MU#)Ȇz??׈^ᘢ#lRm4吺7FǐoFV0cDY:2ΈCMBD>-||xLwEB1dMAPe>|fWmI{_&:"sAyxOy;p-s~;^ӯ=-Tg89 S;:ci4=جYK~vU ɔ}!-N!y˨ɦB5(hUIN8fgc@eO֨f،rtY~W,B1VTasXW ڸ^:נ&[M8#,jUHtt"g[OSj=)g:D_]G f_G\xD8HQDb{W B21(xL$+o3G~|EH{I֊zUA|) S_b[w_;VRY ҘDFLX.G#i(WY+T Gj{'+RkkgE4TVZUizEhzpmR"̓R_ҠI  h5IJ@%F80u:6{ nR:]lTi(af\ȢQ;7|B'P[LHBEVBy)6Ccy^*?[_l:]lPj:FE*ȺO@2]4 QejR S>CPz-( \V.l`/nr̲"55*@ <<~y j1|G/P@:ѹzh̳+D|=@B$A o:%dr:bOOi EWh͓=i&uzҰSz90Nrrbƞh+"  ^'O i+VRUY!QQÈT :/|٨#[/H mgv>H$BeсU%N^ƹ%DŽs:Xaط9M!IV$yqȞG {A#:v8ȉ~5?._9sGgW!^GeW[IR2= %ܭ8^RXM;Rhz~ƾIK-0{MfR1Y WPMO֢ζd3R犠:Z5~|}G}yBzK[s4)E1vwNMyV!I7ѓ\7U $͕RxaE#ҐIF <x \V/p&\VΏeP_i7\mZnԲTRj*![{Kl!%G?gQ|1%G,Rj^H.rnHLWcVٗ9k5EʾHp2Wzqͧ'4)(Zڶ+&z@(ɫ{}{$`vT利 ẓUD^{y U- ` #JrUF/6{yq:L Ch^\,KLux3@ {pBc/#(֘Bm(Io^_~<^,ă{S,ԧ;˰Գe$q@.T/@Kfu/s/2njnמHO!x==)5i T42b %Z~ާP{z+tk ?Qe#K$saz`ⲱ\##I!.㏁l"Kdu$wW|o5x|K~+Yi-ȍqDxѦe}Rc=!k/0eZ 0ɌzUF|A0`$@q V2Y!!?1quhC$|LiS;;?$Ҹ3ؤwo˽TyԄI<Nj e>7rf uۑ7g ozJiV@QjetymAqkFEH 5hm,A<6ʗ÷dtڒZ5DeU·$Aa!`bptV7>b,1yLo (b$8CKa ?m>ESkw'{_r9;gl(?ظ>w&xWt1ELmi E#lD1d޺3L($$I i@rCs/%65m~tj::I&7r?էP3,19׋,p"ELMTa `nRGCZ] 0לl+- aGj9Aiںt?#8Zf| _ش\/(qYΆ11咵Yu[N^H # V$@$e< Vjnjt1LrKv" AVhas!Zx)˱=Uy9Z/6Ab0 2mssh&4Xal!0/B@>,A'dҟ*?F_F(q3_8?r74jP6(ĠoǸstqJ}{l(Lv[138H\Ky\l~e$f d&)JVq0TOЦmCf7Ky|"\=q$6=a|1ƈZl[Fg,NF.b$db88HH@qq$JYĄ5piD|% ҏp-Fn8B>V/J ,7imDFp"0 `-12Mn(Tށ e*3}gm\N$ߓ s]莽\-}~nhS}AaLS|arXU>8Jק[1KG>a2:WF6Νtv=Up p<_(l&\R9OKG˽Sv\zc\@v-U+ʹf!_f*U2+~{",n]seH>nBWuNKi#Jyƍ)N \­W29 U׍5oˑNP9 je|J ̴mcHQ#>e"‘e X(b<>]z|OW\Ğ>n>}Z?!rs3Bڷ}oRb#:X*&  Ma=#jK05G-,o\E-nukQq3";y-޼C9A5Զvc#ޓa}te>lɳ5kIH#7n{6g@M:g vk]u:pBWJ\:[9~14?^MB[ٷP}_ANB^wR/][wziM'O %hn$2fTF@ɤ5%`bc'"*Th6 a]L'q O\q dv܄!`vmq{] LJǴG}L=}?;?<8uߝ]}yT)SUO] 2jzt[(Ҁ0崲B .px^a!p.q},h׽J{`XA:8*iH1ʿ%as?A'>h'aUP&ڞyDOAvIsM+v M?^BۛV;lfYKm+@:Z 20^;D"%0YpScRHRG$aR1ӽSOۋ{MY nUS_wW07oZW` PNno5ׂo}۾rWD}4y#kKcu0D1S5Mkp?ʀCXd/L]L$2g ǝuw;_LF᫼~g>$?=a?m}qG}7;2__؁keٓ@c % λoNpBrPD釽i}I[!+ sBsLT+ ƏeIYۨ |݉y~M!, 6gRQY)8xg~jzp8 o~@(,šM"[+fET" b&v$?ƒqx?$ྯ6*BM{|SQNԟTuC;fƤbNpDi“$qJEc0M,JhDxBI̍Gq$ˆW fsVs/It7d7UgoLwO]'`!Fz\ h5 ObœT CB}$do,G7SEblv|YoN~/x;< , 4-Ap?~趯`p?1FWqqף{a1&p$10y$p<7+uo/Ӎ{Q mZ-A P.kL_-˃BM0)4 B hYН\8^iS|GWOjl"Y[3򵴯pWOJ } ?D6`z"&'Ǟoŭ?K]aR6n.nɄ{S~Km0)b*|s~lݴi1ʒb Ag.݉U'J$&X941Dp!'-Z$8*`ֵ>kk92V0WPiLZeK@A +Z)JSsg^_7%|?co22.i. yj ڷ0[>M](K^ 3C"_>0ʮ|j61>wO\ՍGO/\v1O_udMY+[(^%8gDp}'q9;X _S, 䩯:KO S@m)Qy,$pbV1q6bX9SI6j} 'Vʼ)^7CE|u~:R;]/1|Re އ>5IY|.dv}kyzO l"ړXD{}N"e?W/9]oТK)1}Ѵe1RG d|-U @Gk]1őKəK:#7tknP2- 3J) Sq}6u>]"F*3iΰ:p nC3szKd|/U$VBF>wgnP3I;m̟=| @G?u_dMe@k?W%Ilj 4*:e*PHVi&G'٨]gB/t-@9z&+-dcyq^DbFK3u4&T=#d9#2P(eY8ǴLS)QJI(OU>c^).YAС%ϲݴlk ^4!+>Y^B|гܨLth2ZnʇaG䘹,zYBI{"A0*IckEДm F!AZӘ HBG~mWHD>SEwqT"W/H덥'\iu6~}9O)Aq{H)d&F(a%9 U>2HpK_: /·_p=3I&`oQ|ڲ43 _u y 8جܹ ʜ}-?aB6VV(F3G˃|gP>mO' #hV>O|.z~`TBi(P{9V i0BC˟DSHe='d෰,|PlɅ-%~bz$/[ݺE]"P%Tqk8b¢DQN&,vYQ-9V=5`Dߚ3A5k>`ab5řA6tSn܊+W>TD\?}&|)D?uYBx݈8Hb͝eNAIq( +)7Q#!5.RT[S$, A%cx\vr֖)[aK"` %V:8"܇JrHF $GZ\}EֹbpՓ^\q0`NLHV l(.uVHFg*6"ёe%XL]\pO[Vr+i ݭ@eBTՌUsRbѠT8* BH]B*_i/|@U^_Oqh o05PhЁ7ǣv2LK_ ICo'<_v"p}pQj$E߈V`9juj_R5I/[VzyWqKL~8_[n.aEyr>)Liŀ4|\̜ej$1tpH@5kfH ITX@ UsƗB0-d * +BeP2H 1CKiook=[N Ǽ-(SP4ďk `(rB n зT# ^B )F~?Z(x=Z(@Pr hy-9[>u'߿KNpྱ:r9\,%N9bI#mj Y`18cmuRd2sRk4J_Q҂zXci5' []kHPgBHDdTn2g|$kgIzT 'Y OW1{U0cEYoVcK9 2?VHYLɶvyw;{tTأ[~!;mV ~KLLMGdl5ux=PQ`Hz1ӞbKT{R^j5+d ?j:XWv*y@H.d5GC/ǰv'~xYn〖B/rX8c%ZK=L2J>ҏ8ĦmnB&q:U%ʣ@3¼Ѕb@R"9[^˞(T>WRyop_L(FbRM{AXYUs\~1^lՄV,C*!/ j||ƥU\VBrLHYrOR^iBU0B: )UR4Ff'D.4+U1Ii8VYz}vBs>`7HZ^Xh*j (HB.ʝPnxRLZzT& Bx =pRqWxl功L$;)ߖ #5:WSkUX^RI*z>&AY=Z\F413}8 ͢&+Ib $H䍑#7IL9gX P?@+xn)cIK#m'IGZQcpV(K.țM|F*!9vFzjNDbeih&nfFh$#譞A3fKoB:6?UC/ a[+O2~k:lө#zf'#6Ig qԵ[-n~p[IZw>5cwkcwX[~":a>ӬI WjkX{8dճ`/Ff +?=ƈI}ƀt[X4u%X"ZY cK?P$/4I}cSbzI}L hV)M,9d.K*1gR,pf}E9k3.rISؽZ-/wzEܑ,kg+kV h0[tGLRt1̦KsBV]HVh['AZ&ʜUi9v -r(xgY&4u%@F}DӲ]cX/>-uvT cV*4UT}(?f"V$?bw2AByW}&GVEo A_f łzpx7;#7FDo$Etyv]BcB@#J 򕶉r;{3VٛP)%Z5́ۜ6.4Xv `֨xk" "䩦dhItd  ;z HB:VCG:GqcmcFbʑ<SнK< Yq]G[-ulƞτFTY!`(8z~T2r5D/͇SA V%r03`BT@)Kq7 z:%4i|pf[‚L.%w4flVr$Ť& ,F"4ЀJc+h鹎5h| PrmFlxʏ9`܀#0}3 ;S8s<>܌GK,p @(Lyg01Ŭܔ u&MKֈ8zlG}|7<%,y1گ굠3 Sl.'nDc۵?M85SS *tN84_*NaĊ `IEfm.NI a4I0ؕx)k7i"K]"hk" 1}ފVbIV!bIǵn'6VƻVc~lX5F-0zh|GLWOFw~Wxg.D(Kb ؈ eS4v)lO'C4S{4udoafP k03|j1xꦘFd0qs7G 8r{^kJmòV26O6v'__5/lfstO5.DNFfch|}>!6~>E;v9x\ls ok3h0nevkcSqH?O HTfg J5Cd[_ 7{8ovw??~?7>ޕ_]D~ڏo amƍK{|m`9+ ʏ}"nc^H۳OWߓq߽h^:\ֵ;=#.эGSODzy qrYl}hG=`(&[әtzDoL_ mĜ=_^v>Zc^1"<=)Fkc}:_~y vX3ѻo>|_M?F[ftGBP-OFݸ+y;*  /"um\jڟ.ϦwB $1gwpug",>܍n]ً[:?-p5$ʽ[}>5djV@ӏm7ɇK}`.Zt d:x_DY &OYY}wQ ?uCj|z;y|mXf2 gBR@RBLDt #cA``2AԆq?sؐq tH.MIIlb(ޏ1,rR$( 0$Lm b-y>J]۾p"8vxWZVmMSq#han_HcU?c],Ƭ#%#&k8W_>/p>Կ9?}귷gKlIha\]]x~syz}kʹz3E ;l)*ڰN.\7޾76ŻiNѐrsvzfFK^^ w8i,YkZ^Œӽ~.y0Orܢqx_:=y+'oUjNG-莍V+=;2~c#> J:@KP֔JiA7iZ`Če%P/cXʡ,{}q^UN*' HBJi;ҽn)ߑ $:qtVϰq-e'?&:gf_WyvdCfT BjiԤjQ"du@Sve2f:USvB6Ɗ⶿=i"yEXbeD(B{0qF})l RhrCם8H09P&n\c Q9B825rp ;OlsxCbWx|!\b$M- 0 a}Fm 3W zQξ! .s!a4eؘҔ$lʑx mRhssI9[lE5o򝍙7y'Mz \\6#5f\\+R $"=&ڥLyr$HIBȖ.Df(9ӻpkz8 'rN2&mMT.J0q.i j2PIܗb0@]T&3d4X݈E2@!\٥:%40\-AQMYPS . [oڝE~!l9Nw-Nk$YIgh mk['}b|[LJ;p`B+ԦDJݱpq)x6[7zf㻖"-)BJ2L ޒRiO),}yG ¨dNg'pZqJW/ R^?QL=lj*G+^z~U!t\kQjh%M^!| Qh"AFpH% dܱzvyi_ň2*%<\&ERM>{+TWF\yRaD+d,{#B^5BP^\yrm埼1JSFm٧),JeKM >cZʀdRb !MK!-eMj9hv'w܆R1aUҁ{Sg4@6 XiGiJ^Eb^\g-12Ձs))%CP#9evkEpA.0ԅ @LTRkqF_Ab5ߏbAEn߹>?6)!rLL*GX $%Ȕ nZ>^~3([!q *Y0=q ,h* L(c,h8P`1BmRjC S%sJdY6DECH8c4tΨ`nW&S:$/o} $\ИDDip̱* 1H %9VFDZ>Yԁ tIֆs 9$vW-l1Eh !"GƑI^Kme螠qT>ȴ/vԟRDi,@"B00KG%9( +uwvp:9bsnt7 #+;I-kSWفx.;M)-HݍzxG )8\/s?PW@qĻ F^Cd\9OTu&t[Ês|v@(6udU> ᯺*'!OUȫ"9`[6\iA1R^eO(ÁRߒE]"[YM1)ux0El^/02[1s-}#!%ujHr._|ۃEV=`ʻvuDbxӷNXYoA! v+Whǹ6ZPg[x˷:أ̳#̚`(;>q/4G)]*+Q"M0IA06c"H9V+3#ڠ舮ȓC6~bΟ2i'xUc) YZ=?ww {h=%[:pj\ 4+?]WM/Tnӝk\ԹCYJNROZKẐCm P*NQY#x3fO%UH 4< !+tUH`}ȥUpPSi9J‘v! YIo%CX^YDN^c&bT)Q*$N3W/ڕ{^99Ty2 K =uǴ{_isNa`3Bj9ÚS1X?0&1GE.0*my#fihQ c[0ELwY֋ύ(.9aGy|Q35UanICi]~ 5?a^Ls E$\(!F" $XP\!cf4hdQvFX]$-}-K^B ^ry2o2\)MQO2S d!gyȨ֧ P!Svqq&1ag0BjmYm ܃ 5@BNozaM_[|t,Yc"7 d>N8$NI(JKw;@-5Ǯ`Fh&$ Qp)>Xҩt> g^&}5X VϦX=k`Ӕe8$7Yz֔kf|!0!ؗ0(1#bDHI mTbZ/FQ$(X 5Hժ";kXn5T$oJ;K[~sV$&MJX%d!F'@J?tslƅ ]5,ߢT@R.sRn۲6)JI[ K[iw4]Kۨ( 9B6Cx wv`{K9Q֍@D]j:92p>j5?| ˪vit>RǷ8q|p!''s!3^l敺a ?~a}9qw y!fAE{ ̢l/ϓw 7MmzAW7r[5mQ!˝-Pӝk\FnQ&E]NU+_4N {8 u>΋t)WtufМDX_!߲k߇h6^@,|U)Hsך`|!7.Zr~4j%a ԁ9=3JP(r@ޅ|(J- E-v1blS,ū]bc?;Gcz@jS"e~cc.>knK AwdݻY?1r5id") {Nz>%׏{2IvOfSeS֐|&ZǦ0@rޭ.>SU[qөDyeޭ MMqHq{7)bc:uQŻ @ٙw_ݚD 6=mݐI ߐk~LJoȴž!D(G+ֳs.[J2ӣ;a)Iꯋz+<?EפjAba-:T](Q )Z*qDHn6A^c?=\gLEy_*< #%H'tL"m܇kqƲŨ/YeA<zPq+eW{Kٮ*!IIv) isKÄkM44#nzlhr1&1D<+r^8- a`gEQ;.)HmƕQ q L酃"NО ߯ "+s'3ک^6(aޤ3TPxun,M( UVq3Iu0+ b\H,0*L*,hIOī;A@nLnĵv#=?/ H sxBeZYRFeIJМI%Nq! ˼2]040[ȧz02:奠2)Z܊2%,9JWQ) TQ*BFa 6V}5750 {{zOF-x7K a L2c4JP)*PHQ),CZCfP QKH@"D&YT+Lu3lE FGQ,ch*~n- f'Z,0:eD$ f4DZ)Tw;sAg=L2r?[.q.$lnu  \@;eV`Z`,d3m0X.1"fV*N^~[1tvzE${V>%%2uخRO}KP㥿?MK7eWfuİvBIq![ Cڦ{w qܙdbV>t<3tHʔNrtA.7%o[.I&%*Lw퍶dt,F D'F1tޮZuzntwXWwS՝ću CyS}.pZ@+uҵn!z;dde (:NޢC$GR_BD nZʆJ6,ˆyOɆ 1Fiȯ:7[0(UNS^~J%+$MA)fʉJmPѦ}xR3"VjH@$Q$T#Ngi/e }Üd;1hܦm{~DibW$RKli(ґTh& 5Z`,xhzFBw FWuB2(7Y۳[6OCdZFfq ZʢPKNKe+TU$Dk@fH$^XbKD2YS ܷBc7LRNp (2kyjB2*E!6,,H fYn`)s^E ~ߢ)WRnVI%Bnv} vLÂjKɢAbUG{МSjh7oS 7>Tk%>:1g8a@b3׺7Ҷ[sv_M1\DOo?%XoJwNjIwcT)DDpN~p^gpr ݊ fse;'`%糘s_&G|s,\$ޭh!_,SҦyɒWиHpYLLs{y iSUi*KH*aB4 `"R为vaf̏'o${͋=;`9; P@0m VoJ=`]|=ΞͨK3xkIgl~[s?t8A Ņ_%yjPy(ޖrXyɠAg(nЧ6wKmwvvgUiS[ܘ;wepY~[vgbCE)/dݫ3?~@+h\,i1նуΐ_?$+1hk'NF7di~avb ܻ^2he)2ZBVa{C]cАso0r3LsK}'o|ϐ's7W4 _8;e` @vaHVz~|^ 9rcFI { T$ҝSzADp8 ҽ= ѽZ1@E}7{xӦmSmD:LGpZGnJwj uaD ByZ Y YS$-'`9})/o~e$#$C$ca4i8qnQqpTH u `D@MMT(0 riLTp 4: P`ZxV+ؓVXԫ; tRJ )eX*u-2%RS(bR+*+Q)dKbBH d2n@BIsޭj\Mf@Rz ؽbgU尪*>Ӫ `F@La2)$ x][tdޛ|JWKZ9l_Tm?Szc6枏OU+3;f*H {Sަ%~lFm8B ý;I~0iZju}Ѳi#H-Q=7pM:$(ģDnuPۛ/5\kւ]:m"KI4@ Bu '^^UƘT2 1{VS=SGi\\)3')y1JqYR8LqZ։j*UU];/)rW?{|ص=utM Azȹ'su(t$q1Ik\EI姶"\uЦX7[d.bZ;4qV-9 icj& !a=WuمOW׺y:MU*ivXW_*iܒ4>E\RNq[܎&TLF=#7 G:|ݶdX[Hcet*"]37dǁ8x<ZB 8eW øHō"438Qy-ylq-#):֌-"|c\c~9% vg\83#P=4H-/@Xq 1hIbT@Q٠kZTɭ>}0j@(g )UuNWxWϦOu[x˲Np]RHfY@K ,RǕn˨"Т8sQ&Ñ(i*ђf2*`*pQ0d& µ( 9AP8JKUo1k<׆V: |uRfzebƮw诳B!pm]wQ]iW2xKa(@v.SUYRR&qnH`d2)eS,3q2)F}pfBC;9czJ5F;?enfBmIYhFl ̯?;VՏ(c~ȇ^X^^?DDbg:&+OՖr|[[<]cklآÃ)_UX*J@zri- ;n&_U'x졜3)$[uS q3kÇ?Z9U ؝ϟn{QQVu5[s\V:zb^3)/V.%no˦?[yGf>b"bL?V='eaȭ6ާgӞ4)R}xHޫ/5y0-Mo ϳr|!ɏ3ckZuyyW,Wϫ;1v^%2^BHo8d(~Krf+ˍ[=иQ<,@ŋdÒ`[f82q,@ǠW‘T|0"LʬZaH˿`LJtibnԼƧJem^ܰ a1"dn Wb{J+\PA4D˲!EN2\  -8 R*t(V`JxiOQY(Q QN] h҈!X|07 swGA!0~W`7mvGV(vONA#>$;dc٠GtwÈABLĹ @DŗGa<O+y:pt0nqnP+>TyNMOצ6OL0;i^KC{XYhUlY7;oKǕ)}~ЕgLA"ItY?ݟկ0ՒAR?C^ឨܫEtRjbCsߋtЃ=QzP>H@zG7)/9ǫ*rYlFC=^sFWe/BM?7B,? G"&LQLWNN!**$g(*ĉ"S!EQq"" ND@V葨I@榫@h`PBA;UW*>JU@~R)\XwysʄV(ncS׽l9Ze1KFrJ"pQ0d& µ(,MMd` E 4qSp__k:grIJdVZg@%eiȊ !@ !N`8$(bZ֚NcSjgΟj{:y L@4gt08{CBD53x5?|8>.NRaP拆 zvyŶ7i{rwe=HzY=R> g{vF!j*Iv`3U%$ڲj3>2# լ7>g?)nˢOwM([uqڻ{xq!6]}8][|_o {i*N#b5FVǑT1lPc JiItP۹PQJaWH4#𮄥IG ^8q$Rցgg^1s|yAʡ@K( (Q4@D / ]JI`Lqx -sfAfca}y^Z#އ2n݋Wp1X߿L@dMw[bqdi+2x= z;U׷}aXIa|lKZ{{n,z)G0ZܛQq1 N=3=09&!hpbJ! a}9'?_u:Uܪ7d1ys#i48#;J#;oXjK%]%+3mZ&A!,^!=2%QGikD7#k@4[8b&@)BD5OH @H1) 9ۈ_q1?FY01Z~ e_3*\:%v}=8 >B:rar\V$jYܼ7GR U`l<Dנּ|B|\1,„QR&D bac"bCGrد (y!7$ !KB"# Fa b v_c$I "";ؔK薂GJn9LSɂ)IU"|3<2x _`lJ,'cdq%vzع 8s.Iv05~^Ra_!m^Ec񤂡@yRJ^foVgknW͓b92tE6XTdu:rDEG/pMp̞*_k^ v0lw<զ{O} Q;a?n! #<`* YE1A` UeD7XpGsm[-GN%=~u%َʱẹz kz yY,z ccCDcoo1>s], PՌ,d<Ǎ,Ae2Ao|Y' k  ֪PaȬL󥎨m ݵKD,wRk&(n۱.@obc}٨Eޓ;JͱBln{S,]PR#ŦPdm8 w#^(ʿ,&p#cH݅4 JfQ>orhAsysk+ >W k8a}P  t./wGy/x5勪AzTQlZn;Щ?6۰4s 6E&+ [T_),9hD9w9OT%]&M2/݅P҂q}eڷF̥ūK\6&?/ #e)FUǶU5one2 0KFH$Un[tܣհ-)[pCŶ wݥ {[rOVTTw}k [@6HU4z9Z`=nb1* r-Ǩ"vV77'[v۷A6w[r@޺q"Ũ@IFa3Ze1_,,|h$ٟBqX}d\q5@{#`l#EK÷iDC,^- ^J3_LucH*,nm3}eJ)` m3{"ͮTkMc IfTFQe%)d%]h0y>dzb;dc߿͗,/ ^g߃|xgF{>ygwG267>JrQwf=nSyCby.ܾRvtj.U܆@k›Mvrk-;4lk^;"o IkW2-uswണQi-$@SNjV&-GtJaX1LiJwA 85nx9634oVSmz_1sȮMBANкq&ha.*,%\X%a<H¤TpQ7~g*f5nGKq $U)${34o9 KN:I3:֝(ʭ|c1iNqT1ZYy )%6>0A2fC@[ CˡBf,Mmj' +ҰӹN7.}jK@-/ҿ߿O|#t޼<&b#3wEo${2 d')\ds"w7bv+GO>%YZT.WQ1ŭ;'mz%Ji7ve~G #`nA[#/>D=t1dںm{&jR %AE!e(YYt:;.D]5r@B:M"TYʕhր !DISFBʵu8+!%LBDŽddIHFv$马xxtsf`. A[k xn>ɑTe#kvkg,-l{#b2yM&v/ :k;'-Y#ݏrY3[bnwdH;hD[j-tDM(ޣY  leK7c!.ihΌ[i_311KTs`,輜GnD9t_`@ְcb*& GOw|k\&Y8% i<`(D@j$Gc?579Hv޺H?VJZXqqRjpN81g%~rLPgk1WNq!%(9_\HxĔDvx-ӀmVwJKu!Bt r'h!@Egzƕ_rgn/a^3HNegѠni}|LEm٦,Qwid:6eX,ɪ27~v_xoxFXta*q>_q*)*oq$ipܸ1kuw ŊB2~-ls@8_Arc _/}ʝEqǧN"/W` ze<c>1L%P'}LO@*x5B}4 x߻ ƝC@uGP_{G>?x> LSmݼ<}<kQ|/\V[~ꊽF܋-aSs}95R4tc4140$ evX ؄Dy0[̟m/ bk+n՛",N%V]F !G²#.&Vy+l8lIJiq%&S*bb"#4SV!82X)zE4&HjLc444U,E CP2"(0<=9FqZBY*wB`L݈]q#jΣְDcaRRD(#$IdaذĈ$i C %0s3S [|ߒTƯN[pnܓ VwGiY /n=#BV'#EaQb85aBH[L$*ejF!#E*&c#("Wצ̜ɦ̘F?V ci$vch:A,ܡv||}>U@F{A2Ee ʼ!"O_< @ՖO'@]fbסl t Eei^'kJ ì%>tc*̀̽q'Fȹ FQo?pb >z# c̦?_t􏷙U0%rnV8] Zn6O/c 2y`m" f|rj9B~Vwωmxl&*-u /6ziglP g0?9wnf}ʺP9{ך+۱xsԼ$Odȱm`T,IH p|Vr=Ym3&s m T.u _dzܑ+AeȚP+|7^- 5Dֹv~0ٱzŏZK-ҶC>q%F$ۮKٚ-,,XK@,&d(Lx&ZD4!RhXGEƘz"bCGE),kQ۾߿-҂b3903@1LlZ g$Wi.NMu?q9`ojKnyA+rl:)@ ȅ)"DJfCqĭ4LVTTq PD Ǜ*Q8(IPG‹+{ur:vgOsA/M, ]J:RN\$ʳ̑. j65w}٢y#_JVnwhxSiw: 9# oBs'̳i['AWrE)7{k |лVnQmagA\]/ .sI6ߑmP;˞(P=װ!;jNR&@A㖨<:5DuCWG[HVHcq֍SZ,9Fw(e0+~~zzJsop`"oC"i`(V #>r֍Z,9:w($wql~(}mΨ|RlBWmMmCiqr}V{qvfO쩣Or~ΎWh M[j>W6Br*Pjm+<9ZxZP UNS[?^ਦWR$E E]K/ ;G)Fi.V-ysԵ)As޴͑k0P,ZNlݒ)Ztwyj:[mn}/˷QˣK{Ws:W^w|twg6[ıͻclJ4y ☣Szqބ}Db U? HH`4_X0v7}WfbZ ^T%/nRlx3D\}.ƝssoisT<"7C cMkJdD"Mr-{+0E&3-7NxnBvgƥͰdXr>S’@U/n^f[!H%r*$ײy\9t]+" "'}čj):EUd2 _D"R7ʸ ϊ_z -D 9r:LvA_b>s}}4 fY/V\Cl>s tjP=TP\ΰP+0{QF9M^KAr9ȵCeGq2A0NJIĘ 22J# |E0JIz y9X&{hץ\L*ILnM7 DN*b&IXi4,B%8D2P:aGqrrN("wI)#$*5uS $ 9 c 0&1|Z"16eD.a U K,3>~<ƺֈ} -T0a:כY 2񗏯ߓL$"1?_,xX$[_/ H260 槙 ~y],W[>5~x~Lr/[l t E Lea=jF6<9 ìñC%U)q0_9VI;͙̾e~)f7,L`w^ݑ<ʓP'C#*$&ᙼEx0W~3y6::lM4Ÿen?FG)6ncFB6J=jn*KɺZ=.B.e!㝉v}-E  q[;E/nz]\]ZWiyKz͔݁.B4mEl4OAגyv8EU-Hsu2v0\QҥdcSZ5N׷ޙ!yAvx<;HShI:k=jGrBl2uQS]2p}CR%i|lW۷R1^KI[7NfݡR7zuqjg&8_ͦ/  M )^߼V)]85[/ N j[d C@L/&h,WhQ繹S^5H8I<88pf$,աx`nB`)װRNcPDMk.JjX*GHa$D-fU ۼ ;VܝTz3R<Øm*սۥʽ9uuLmgi>%#ïQFǼTƈCKd~8X4_?) c{2 }Z`z5q;6߹]ve 䌼V⧗e{ O_X k(iϒ0R-DA89>`ZYB$j>i.2O'Iu9Kd _6_C)̃t1{ fݭK;L%a fRQfF,tYJ"[5a^Xn#3`.ŭBLc͐E&j!1t+"{yrV@ e&o0t+s8qHXRP+։'Gt[НLg!Ta҅>'D"!dz{8ʉEjnpx!s)}rt~tS\͠, \(&IRgBΔ>rIq\ !\Kڨ-@}R<({2F!ޫ=O ;8z։XXB)^uwNϢ}|K[uRdϢ?~ 5g}sL\tYY=KG[c/xnc.{CU3D[VK?֦:;3N/529V^PL/:pBH+BreT'!k~jmSCԎ2C(u(FMqaӫIq#q&yMJH"hVF!pID(\AӔKFHj}%nSho,-c1^{{:x S j ìcRDRH :v%m7Lr.U VFk85O֋$/,n|4F`-cw> (в?!h@¸^~g~Ov>ّdGS~n?!U<#DISNY7b6&Um!N~mZRϭ.X va9heaNm\rd:EGx[ez_Ie0Hhblt'/ yV 4Rۍ%Sudf6і`04uãklub\z|+U\_oj)lA34p: 'K :l:q[eݕq)F8m0R0Y'L>0%gӑ#]u1ofd3e>u䚯t78z ^$f O]re؈&-"C[͟K}'j3i!Z끥D_Xn]3Ok4%^皰DIuC > {1NF]MNr iq؅(|材_IYfw:_|fx/ }y8 q%K) iil&x'4[Ue"D3{-n*ѱ|yL!0(@D=n(D U:8Q' I82yA u6)f~:+7Ê~ ˍb_3#F6(usZXX,q?L]T,atQϢNwSN )SmIH"3̼:`b_\26WbwUeD'swe:FﮩkfwX1"p2_}J.u2k[W>v(`Bzi€7FSdS&ɀqIZ!9tht'YJl'kKu'a"L⳧z 8 (QN0 1aT Œ8A(|8>#@OqasasƝ%x6d:F!4Ӣ/A$y""W?0։7|?`ڬ7:Yy<`A@H ~/32 "@"OCPF0d‡!qD EJ}B,Gch$Hf jSZLF:펞 ZM"iEPZ\ӯ?v3RB|Gqh~ag_>~lo g0PeL7|Vxr?PaC|f/ب6}YMl'`{{FR&GJ}SxG@ďDGr„l݄|GgzAHƩpsQ$g$+q̨5;bOflV9Jjgr<.yb<}0~~κy?Zpjʬ=^1l7Kyw{?/ԇwzܝspL0Hn`Yґ{w!.̝W(B*d@]G6C^D_8/U+e-%HʻTlfya@pI܉. u;&иjpjMBv_B ]5"/z)l''-MFv`-}vtI[)m tq A[v_ߊ#pZT qၝ\:vqg@K1jV rJB"k*ܔ+%fRyw׭mvk*W6 k,bN)Za;\pCR]xCI<REVH=4s(=jGRA'uluo%z?j2{)VzoLЎg u28^2sÊ"I J f0%gu3Dhy}/0d:mzsa䚯tNop9 DI >3PE̿?ϗ2ˍЦ 8B[.nk5+d8I/3uaS%jz&ozYk9Edn5-g 6B"k+r sX9E0 dHPiH JD9$R-JTE= QV~ĭK4 1Uĭє[n*gtՉ# c0iPUԭ̪r6]bV?ucdF&.8k UV2q2U"e`n6Ѵh@XBWwc1&JQ ʔ!5V8ry+R Ze/-BQ<{hc܉gUea^tS.8}Vfw1Ct$^ PA]rһܭb)ΘS ֟u_**oLjc{`A}X~MW.NsG|8;Uvy^-,1? Bֳw X3U52o@K3!v*Hdk_R3~NFmgUF$Nmc{v|7j 7Idikj,7@"(nN2*(1-h.:{ȸ<{i٧JqNrdBB@2C H(G'pA `D8Qbek7wG$d02TL,HaG%HD@~O IA RrLba Ψ;wY/d˰ԩ qʺ^8}E=LX:LW@Q:D} 84AD&a tֹ(~ C($`@ŕT fX<<H@MƹAvw'igd{FqߍU9CԨ` hڲ N0Q8 JM) $HRv2tJI95R*X3#u5)tkϴ^h6)z6hޫTV=!L@?Fe8[d{@l4V 0PaL7|V[5K[8x|LeZEm`NU{-94'bH? G`X) w? _?-g~`R}BJ,ueӺ#XOx8ZHBll:MޯT[ySk]0Suãlu&ږzǓ"ftJj k?)2_Ap)^c9Ec;l-ScE Akԫt6Id̢ KQ؃ws=ݯ2U` fnBv}.+4V!0C Q$8<aD1af5&*ąn7ŒJcޚdgzmYR6߅ZQ EY~(LYBTLL(.>Xt6}df$Vu/~BM9ܳ/VTYfvaPvLz>Nyd'Tpb<)͖)0q~0ҧ @4!>DjICD0o-'{=WSM?D"hm1p&o9ۉ.R5v42}z^3ޭhC6B>~8zrexXpW 1e9 "HV :N zS8qꃘKN/Gҗ"ABT=dQ$ G$L$q}!aہJ(V}5H$8bpB#$Pq$DL`H1M@$Rswe:KQH͢BES+h<{࠰\g ( b9 @t!'(oF'']F$Ђ8--wtJc[MI%v tg>0oJ'aF|'I{4-]8^BE9riftpw_<ξO_qt*4>A`b>s:=~Ge<]wx_8ßTK+ax,5:wMǻBODʐ9Z"[7Ny11CfF4X/}}E뀣7GA"!sgpX=t%NYFB"áaKi5u x=`4s B5FReS0"kZZuD Eg44$hޱ$RhbQ- B"FJIefٓnޅ'-G]jM_Q*Y-9Q Qn8?Gѭ&P/\Og7_[]ٳM-}!]Mj[pAxAASh~K{]܏2'|e&y .>D'U,'oa8xF jo>1ʣfBÇ_b5M?C-eъ jdoA`mqTӎ# d aTSϔ9h0)9̞!, Nyl 86In4̔mkFO!sϽD >\~ctJt)7o#26M2dV=! +T0$Vzysг7Sl-f/ڒws6JJʆp9$`T?CcZ{ܸ_"7+@03768ratl6ͬ]o[3j,[R:b[yX"qGD(OFwTȢGiuU<ѭ ;CZ'MC=At,xfqۣm;8'j1 _vF8p[Ă;BȒTmn߼V(^̈SM\Ô]2Yujr֗@6*\qZy5\JZ=+%BD'd*+廻0a v+]H3cP2wᆓqHcYpV{IN cIX ҨQz:C.P)Q.3x*Hd Ěd>>X#֔IkH4 bEfTN1%]̀I}LiܑnyyV';Қm\(]mmYj~Vכ%5sӯa1hD̲\߯&k`f>S l>Y'MJ+[߾Q&f@Mh;Ytc$Ӓ41͘2.dzI 7\Ig"6*4r&rcJ|"C8N ZPDžrl;iV1 %$b\8 7Eq x eM0Œ[ KL`+X0 K/d( HFܟ_ R<62>0*]O+fb#h/na=}@pPXLcWD2"qQ.z%Znkm@qʸ84$QRftif.N+YRe<+ $lL#\!17;d$d ^dFIvH1"\vA<&Kbq;%|E,KocS`议eV)-ܗ"Ƿ/i`~O^~]0Oq;sO o7HlXq֔ %7W+B'$/^D+l?/S7pOVJS([=xsFKOz5-ЇqPɑ&`n $\Xp<ϧ^PjfLdu. |-%Mxb|ܭv" H5) (ߡ2\A$Aumx5csf<8hiHuR Lq0էU:kegOCLG 㼉Ra[ڳݳ E!V%}):kO#еǑz 1;kcki?n3'q=ܡ.\ RO0!d4mO"u\d1b$#1"keR#(1;%*UQ9*UMppZ;u˄ ImHd &`|ra 1ҙ(N w^K45;+axޡ>{ؾx??d_ï3VW]f(@dM^o.]=ڿ[o&fFBp.jfQ4ьGr0#iVD)S?r* $Qjǟ[@ H9'@*rܢ\~3U!  %dd=_94u"fJ)cgHxmD*2I"4"pLw0ґ4,%Aw |V$<| 6=DPRlG) 9C0smn=/024;S#EiLž+a0hN %#+|){ z~1گ8# y쾹\: HsךZo.n\bY+iY{9}pdHA8+c ׁ|ɐiEA[m" Bx_ژ]@+㑄4 u,~t뺱YX8Uz9sx0`qx͸RMUObejٖ5hf7ᬗNސ-y 0px;ima^1 ޘF cJa.*UZp۞ VL GE*˻ TGhvy?QJ$.n3=keB6pp tW&Oi_ {ck4Y•Uؒ ٪Cdnխ_czU´엫n_nR1`q\^Ƴ]8ȭsIX]qWnL^Օ[M4ǒk/|++Ժlk$V?.Zyr iNI-;UX}C$C0ldCɛ"GeiŇh1 >y[cGII:nU1:UQDuh?9өiV}HvcŇh1>ysG䧠ݪb2tvrQlhڭ6@VcJ#,5=HV`c h?v)Uv- &K鶤\:B7r|_mԒJr9R{KBtr:uW>Z DQdG}4<]HE<[>flp'7߬^`ެ޽-E;taZ(Ys)1wV\/B!'APW<@f1S5s*' 4g}2 N}հd)G}NrҰ IE꼋1'~fI=EiNwV$냞"R"*zo(VQG\ yWAvZ+ Q"jQ:UH , -|iOW~&]{n]7͇)OѼn4;M[^S-0Mn^J>Bѹӎ.H4d}[t3JPQy_iy=ԚqT7l358mhᭋ;[B+XC^':N&m+Zy+~##9͵ l2^јŃI=>n" MBCQMдQ(p]mxBBp=FqֶfQ'juZ9]LT Q/rEQ9_qA8&h9%Fqk$b阉le &,K/vNBQ$ 3zS[g+J$ǗqIq*scO6.(BdطU҈h*=3Q#̨ʓ9z]NsXi{Xrc4Rd&y& e(>D[\vV#?VAeT5}ܾށfǟi2.ڿ\2j)R*rv)Ht.;W˕BƬ !҉=\]I׿[j@T.dJ'Qi7x+!PRR{hPd%S%2r';q.8 }ɳB9ʚӧ3sp}f6EjB }C+_q|xw]\,:Kǎә=MڲȹrXHC#:t!K(jPC |xǩ;C U RGWyAD,wviE%J9B'CO V{ih6^(Q^CU(-(҄9G?TB9J䡗0Թ`Օ8aӜ,K(wt9#R3|[q*4W !RzoL&2Knlz^N3o ~SpWodIi&ߦ~0/J]и.Z7\i6)&z4ɥ/T50V(͔0>7w{vGM|kQh=K g a."!ɜAM$ױp:_%!cWFK݊N <$Ƞ!q<ԉaBLpɱ,fGd y[kq%9_"M h=ڪ XK!NNdmeF*K!N%p9)Mp!$qVeUP\ Ym*}Ta^ imna8sG_ J-v l. fYL:RjsgU ) *|͟KaPm?ϠnӥIҤG TCKzNr" i^KGuEp>NRGAqG@aI](˒MR(ʈDY2J3]Q Ƭ>D]}p_ kT̨>0Ttg# !Ԅ_f0\HK3BmޱY;7N)udb.ŎD7"/NZi"h}feiR]R`G TIeG[ eL$M =MϾx=:U>i ͚#.J[SIPZfPdʼn#0DXQn&bNJxPHDt5pOݠJht7TF  x{U~*?!-$G $8&aN'$iiNX00FN0-1mⒸ"R#g$ƘE1wq v41(%WddD_U;@R26/?"̥9 /}f r؃ SJU4iyP5ˆ \ kK|=׃x<]8 ?l)]F" KREFDDs+$X%`G ,UI)2v`Tj%DksΒ }祘6ɺ;Ł(f:3*KTF["*AtWE (Y)&Jjcc8.4s~~KaD*=H dR+$?`fzzP ]ش"N%1[EQ.09D*Ic}rHW:Z)xFI|M͗Y}^zXlY]= WZ/Eo_v|a+ߓ߯rG܂em0a""Xm~i+N-'ЎnH?OI4_V=O+~^NpoxG/ z[ c/ -yx<jTJ*#:B &.c Iˤs !&/a l# ϊ4D:%rYebIB8Qäр8v0 , Fa#MZqb~";㙧EAݕ:XQZc b`R c؞iBbBL}l[PZs䅘3!Ġc ӗ,h5ZUjDiJfM_N\4Y<|i0t-}XmnOG*;Pm5[@4ڃJxP ZmgtȺ|V܇>Nu`n-yxDO0f\~6~WЩ][s6+*d'qt2n&]Iz hk#KZINkk,AioVtR";88Υn5ZMNjwks6Џ8oPu:oEPw gpԴW C? N]C̻r<[u4_l2σtn~s{tHʎ<*<*B"Fw,e4engWYB[jLV- ri8 '?`~Ǽ!YH|xިUKe$gYG/խw?WOFgyyɈeO45=lA"el5!nfHXտ1rq)v@3߭SAl4+Sd=[nWcy}Aoz] NwE@# TG1@}Yx^rJ7~ï?:): 03qng*kx^cmÑQNmߍix}u:f6&j$s6ɾQnRqcқ{zAPr7<uܝ[RoqT,~+_LXH1 +-z<T.._M=HOeC܃V#|[wkM[޲!⤇ABi꼇l=CH#6HE{=lGC]{O_8@ ` 7OSk{Ըݹ`8uVBI65*%yt5Rw4M<RȞ). p5|je! 0ppNڵG!^߇@1\PJqO K,.eFs\n4:WQـX8U\*B^vBC-{5"[d%LGꚶU0NV/rD] l˷!uQ?{0  h^KpZ֛8pj`JpwN/Ý֝Z܍F*2 ocf5ҀMS--MIk+!.ÉaO-VZVPșxT4WNCLJp\t*$6I7w BA:z81"Nݺ߶ŀ7RiP0xPn:YGĀ*Pʻ bSF4•G3A-E3H(%YAV}EU^]Zb}h?I=-~W7$ZOR+'g;^&ROYkaD_>+ ) iEt3ffBGЀi4Jk] /2}l=2:B%NI>Ӱ%TAI!P`f>{)VIA"PkcIJS:\F$x7LI{:Hi 6hz%F+ {?妩(vOf*1ݚ-ZkldNnSoeVtˉq߯{0[l~#`>ǜ_rfAD0[[J!0- 5ڹXa,4f3WP&tqWXEC4%-'Z\^FtI5IHdDTy3!Rs^' T7N^i3!.o:i N=N՞h~0ϊ Y ?lNhguw*̱NAk#ުe ,υ8A:&zԠ;_]τ&عqT=3 jW kd;vmb($2w^ IZbb7ős12sz>gIy,]:p&'˻ad[Sbf7DN}uFZ3.~ :_.7"UKr ɔx ;ZQ $t-N&eqee1!ljH+r;6عUrP djm=d@ʊrOz2hgvd 2r9*k&"Pg4]D\MFLk90|8 8AcIv1Cl=#%-H")?g$LI P@@qXX1hLQ$0DA0Q!b Հ8S"*w#H"$1RP$BD& .CaxD94XS/MZy{-˚K4=W֣9۫ԇV/E^vu@GܮTW?O mo@3+=V͖zg͕[8Nx3 ֛Ono:X7bs onH+FCw8=ʮ&I"GM$07j91ZHBri1}WV޵A&y~&qe{^ލ~I?ijMLg+5ZԨ1l賖c֋n{GG 5Lm#e5zZ UPSAKsf^5Jꭦzp$ t_CA.]: `Z :>ǪAB.ZF~ b h_aC]»h-UV"@{F֨dI%xm7$ځf~]~(ܽs 8^D <主jg#׭;'"(HQ{|id|y{ u }kܓNʸ%:ٗYܺsB;Tfa6X,-R?7q4AK"Xc1pځf 沥U(xCPw-*<z*8'R;NWW m!AC` =U`8u"N6l*4XzCӫU`,O)VÞNRt(F(⼩ߩw[kR-" ~T97͇pdoѽ Dm.(ړyAC.=X yh4aNe <1 (5kdap|/}X:wt^`䇘X/\+WH% ~$!$³yX6˙H6D+_a&;{,B3A|Cў4ϛ`f|bf=c__{͜VOi?>vg-?z=}?鵱AbEt3ff?ꐎ:/AiO/&0@}0XUj@|{LI{:d(['M}txA"}*Ghaïj5M]j1_aΫD71`c t$RD/0Za.ﱒ#w&^..+;ZOO?s?Sq"gB;y EFMN7ӗɱ{H*-Hh5_nirE⃵U8`0t\N.s {!)lHeƉDn1pM:|rºU dQ1Eq#E P 80NXȰ *NP zAU8bU);vd _4큡֦Mbx_EqB XF(B1r.q2`(qD1ИNd䤷HfZJbıDQ̵;9W1! #^ lF3/FӘM*g' 'P8T!#S(EC RU 4 PL RBU@7p#Kkp3 "nZZ#5N8N 008_B0gSFJn?N(@3'6Z$n%9"( C(RR$Njd>TՕ˼Ahcccȋ(As9nqs/yƽ4yE#'=-dZ!`{h AZ?uzXN{UU/Z7=Rݏw >.7~V1y:ovklcy{~ʹa,@L"9$qҖD76 挅G+0]oVƩ^h5/]_I (TX) V 1S+b@/d!2QZDZ?p c:CqDDOԢiUs#A#6(o-a3M):tu(ga5hh/lc{'K3m1o6Zśɰ%0㮪P}mtJrj\]uCU☀/yKYY,FC闪0wVa̿܊Qk9t?#! ]vX\ NEw=Av'o 9] g LXSbB 4 ULIL1IHC ebL&)C OiQpS>ʢYP :I!D`AaD "$"QA5JbpB!}z1]oEWDxFN7[ALV+~ïS)Rh@B_g0g' ^b0HǂP( _R>.dxmOGTrZ\Gf݌{4B֑=Z<,T ||.>=A;$ޙ N?һf8 I2Rs4ܧXͳֿqFdL/1ZۆLz2Q2&XTf~ Iȟ\D}dʾq{$j|1pwTn=nXԛδ@]['Q"P-c&УvEt|QEtݼݚ?d ;%;rNc֩cFng*0-Vgsξm8JS\dl^{8(7 Ą Tf7j5^~]k! /_]FSG R_Mr_!A¬hǾyPB%Vt3G܋!'M AI=GTU$;a2S3weݿBf d`LˋOk<5ktw+%u$YRU+$uvfHFÃLUv#׏rYٟu[[m97 ֓^egrZdRl|k*53%1Z*R5mUH@PlPZ  ]"H5FFsv2GGEjywEBfiT} u%${`}]ܧ43D|ܱ'*xȖ9%$߆$hue1%+//@C(C9-`B)T6 {{i|LRp?%msC@$Ch6UMq24[ƕp%.Nֶ]kb 6%ڻa !D4ֵeflIA-O r];kA{߼f8ktýU^ ;&lu6i!f[''_Ӿ+-di镝_RK˂jRvRUjf['+m$EiүevFajFcLQ N2?@C|R&ˉenI-͛B$$; 9z(R-yj]cEfnEЕ9j,[VT\[ o=L2tעwlߤQ{gQ{}w [vQsv/7r.oQh mK-ܜܗ|I9=[ b@EkTkb='띧_0X"vn ;| 6?c#MV Ib; 9;?e0@IUZɾ:% DԶ#[>XZ喦RKLfcJKc)m9vs'6#`boԜ8k&*Z,6 Ge88DX !glΏݧ8 EcE6  kFS[Y?M O>vXMi}fCUR5֊sSWi%ӜU1 vQI;Tniqmi{Zbt!]]|C!\?7xyj+Nq0y@_񦕋t $x|xqqȲZd1`M^䆗6ޣi%;-/eaܯz`3*~i=7D3*et+6aI@d BT6|[n}e|v}2Ԁ\}Mk=6Di oC?ֺ1NkTH=6c)bټF;&6>xJd'XysVtTq?mRKLC9v\f8:қVd`ж1&c1 26ݡh[rJ_ٌ͉gL!7>H2Ja˵q ή83iANI#fN 4 EF @\O_UbA3QĄ 9'###ɜ1.WmPmV4$?~nЉ4';e,rn8LF, wWZ@MOޥS4! e`Z ZG;HT[A,l٬DA21:M9җXch]2Y't֛Vc(À& fF +Ď10L@̬Jii0q FrNF^&̄iDZC*+BCImȋ#6l.,X[ hQ]ÑjRE,}huQ93@]byԉ^Ç׻4t;ăt ,JTOcV qg~Y)˱lF[q9<_K(%m=׽cs 2HƨM:nY3ٞnI)v]Fw/Pn frI`E>5 GM+7@;t5V!*Wíflecx--6˽{]N72kco5q]'@ &X-7 ?ޙ(8܆H*ScǦm9Il̐RbJfL*)ZŮ 9pjk VM·U n2G&V:a^jľNn'5VXWl)ׅn߇>X@=]BWYtZzezm׃ȱvYJ`ƹacY$o<;vuKYz T,SMwlAi), [tAvj۸ZlE!h+T:-]Pm`]R]iKڞS}Lu%1ijoÈs39A{{eaNK8J7]C.KVJJ7t|grٚ7̤ `׸< 7_V8VFf7EV G/=N!48M/ǔP5; ss7n`>fo$fV| 3<$Uc>gu\z+Ib*%=\'bSgc>FFlHlQQLΏ+$#R<&}ZBǠuG<G68M ^笠^c}a*cvzmqu/_o/(ʪAKCΜK/<1ݝ76{MC ^ZVR-gB^β~^#|{dM% %DYτf;|2D%ue&T6 z"ZE B]r!&j-#jrk}r !j!`ݯئ EASBHZI([덍eUۚ([_lDl[YF;&fBL/[9JTq+ o[._?z (3l#Ĩ7[n! =wwN?GWFWg)tb:|ޟNNOGWO77tOvӒ{%WlwzvCbYfęe6ɺW|~v1&!qޢ2D9 r%rL1u/]pW"_z>D GY㠍`0^4b8Zk` w-`{Hv4 ߒRܽiZrX*? nsJWUet~kW\ ߚwyxfoN-pSvňQяn>=Swnd?pDteiHJZl)%6$+,yB+]LwIho:~Z Z[gRݹ_6Z;ŦtH{6: 3IWY#',h)=<|,Ng_xx_#fn*5/U؄ي/(֧2Ux D|堧X]Ꞿ.չrPõ堎AE7^lG]+hxH`+H5\y!ff=X`c0dwWHgOvnO>ITC@.YB5ww/zs Kpu`j)m_<_^]tk+{[Bu,NϋY`_N/}Qu"(vÓ^}P rYOn+? 2x.bsq=vI||`U1J\^? 7"ԛ?t;*}@>]Dyѩ8UtNmqj6uKY{}7m݆_qnch+W%r:ٓtwlIE~192.168.126.11:17697: read: connection reset by peer" start-of-body= Jan 26 16:58:50 crc kubenswrapper[4856]: I0126 16:58:50.847432 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:40202->192.168.126.11:17697: read: connection reset by peer" Jan 26 16:58:50 crc kubenswrapper[4856]: I0126 16:58:50.847860 4856 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Jan 26 16:58:50 crc kubenswrapper[4856]: I0126 16:58:50.847909 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Jan 26 16:58:51 crc kubenswrapper[4856]: I0126 16:58:51.179478 4856 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Jan 26 16:58:51 crc kubenswrapper[4856]: I0126 16:58:51.229938 4856 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 16:58:51 crc kubenswrapper[4856]: I0126 16:58:51.230217 4856 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:58:51 crc kubenswrapper[4856]: I0126 16:58:51.232129 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:58:51 crc kubenswrapper[4856]: I0126 16:58:51.232184 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:58:51 crc kubenswrapper[4856]: I0126 16:58:51.232198 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:58:51 crc kubenswrapper[4856]: I0126 16:58:51.235399 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 16:58:51 crc kubenswrapper[4856]: I0126 16:58:51.363464 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 04:42:42.097215514 +0000 UTC Jan 26 16:58:51 crc kubenswrapper[4856]: I0126 16:58:51.662132 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Jan 26 16:58:51 crc kubenswrapper[4856]: I0126 16:58:51.662716 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 26 16:58:51 crc kubenswrapper[4856]: I0126 16:58:51.664855 4856 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3f07438e20bdf71c752bb661084c835341999c561bb5442d75e177223881276f" exitCode=255 Jan 26 16:58:51 crc kubenswrapper[4856]: I0126 16:58:51.664952 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"3f07438e20bdf71c752bb661084c835341999c561bb5442d75e177223881276f"} Jan 26 16:58:51 crc kubenswrapper[4856]: I0126 16:58:51.664973 4856 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:58:51 crc kubenswrapper[4856]: I0126 16:58:51.665070 4856 scope.go:117] "RemoveContainer" containerID="9fc5bd8ccf4d2f104d1ef654e18a5851e0cd141cea5247a692b0bdf92c390b4f" Jan 26 16:58:51 crc kubenswrapper[4856]: I0126 16:58:51.665571 4856 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:58:51 crc kubenswrapper[4856]: I0126 16:58:51.665999 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:58:51 crc kubenswrapper[4856]: I0126 16:58:51.666058 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:58:51 crc kubenswrapper[4856]: I0126 16:58:51.666072 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:58:51 crc kubenswrapper[4856]: I0126 16:58:51.666889 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:58:51 crc kubenswrapper[4856]: I0126 16:58:51.666928 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:58:51 crc kubenswrapper[4856]: I0126 16:58:51.666942 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:58:51 crc kubenswrapper[4856]: I0126 16:58:51.667702 4856 scope.go:117] "RemoveContainer" containerID="3f07438e20bdf71c752bb661084c835341999c561bb5442d75e177223881276f" Jan 26 16:58:51 crc kubenswrapper[4856]: E0126 16:58:51.667930 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Jan 26 16:58:52 crc kubenswrapper[4856]: I0126 16:58:52.363849 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 05:25:06.758196884 +0000 UTC Jan 26 16:58:52 crc kubenswrapper[4856]: I0126 16:58:52.660177 4856 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Jan 26 16:58:52 crc kubenswrapper[4856]: I0126 16:58:52.670750 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Jan 26 16:58:53 crc kubenswrapper[4856]: I0126 16:58:53.402081 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-08 05:10:39.45398496 +0000 UTC Jan 26 16:58:53 crc kubenswrapper[4856]: I0126 16:58:53.415125 4856 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 16:58:53 crc kubenswrapper[4856]: I0126 16:58:53.415377 4856 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:58:53 crc kubenswrapper[4856]: I0126 16:58:53.416609 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:58:53 crc kubenswrapper[4856]: I0126 16:58:53.416648 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:58:53 crc kubenswrapper[4856]: I0126 16:58:53.416662 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:58:53 crc kubenswrapper[4856]: I0126 16:58:53.417412 4856 scope.go:117] "RemoveContainer" containerID="3f07438e20bdf71c752bb661084c835341999c561bb5442d75e177223881276f" Jan 26 16:58:53 crc kubenswrapper[4856]: E0126 16:58:53.417658 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Jan 26 16:58:53 crc kubenswrapper[4856]: I0126 16:58:53.518074 4856 csr.go:261] certificate signing request csr-v4mqc is approved, waiting to be issued Jan 26 16:58:53 crc kubenswrapper[4856]: I0126 16:58:53.526238 4856 csr.go:257] certificate signing request csr-v4mqc is issued Jan 26 16:58:53 crc kubenswrapper[4856]: I0126 16:58:53.561825 4856 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Jan 26 16:58:53 crc kubenswrapper[4856]: I0126 16:58:53.795607 4856 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.293913 4856 apiserver.go:52] "Watching apiserver" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.457562 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 03:27:12.045590418 +0000 UTC Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.506463 4856 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.507073 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-dns/node-resolver-t4fq2","openshift-image-registry/node-ca-tp5hk","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c"] Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.508231 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-tp5hk" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.508832 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.509268 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:58:54 crc kubenswrapper[4856]: E0126 16:58:54.509327 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.509723 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:58:54 crc kubenswrapper[4856]: E0126 16:58:54.509776 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.509867 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.510347 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.510782 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:58:54 crc kubenswrapper[4856]: E0126 16:58:54.510834 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.510895 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-t4fq2" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.547636 4856 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-01-26 16:53:53 +0000 UTC, rotation deadline is 2026-11-12 20:49:32.130259927 +0000 UTC Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.547695 4856 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6963h50m37.582568216s for next certificate rotation Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.571868 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.572042 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.572215 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.572295 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.572042 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.572542 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.572569 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.572617 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.572799 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.575746 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.575943 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.576171 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.576397 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.576614 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.581179 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.582890 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.590307 4856 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.600886 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.605506 4856 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.616480 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.627299 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.637370 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.651704 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.658599 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.658648 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.658681 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.658708 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.658749 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.658766 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.658784 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.658798 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.658818 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.658833 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.658927 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.658946 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.658986 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659006 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659023 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659045 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659043 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659099 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659044 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659159 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659178 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659213 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659229 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659240 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659253 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659296 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659313 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659329 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659344 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659376 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659392 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659407 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659422 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659455 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659476 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659491 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659501 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659507 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659597 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659640 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659657 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659677 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659696 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659691 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659745 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659762 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659746 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659781 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659798 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659816 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659833 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659875 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659893 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659910 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659931 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659946 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659962 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659977 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659992 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660007 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660022 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660047 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660063 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660078 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660095 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660117 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660133 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660153 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660170 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660194 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660216 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660254 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660273 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660288 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660305 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660325 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660359 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660385 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660403 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660420 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660441 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660463 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660505 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660550 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660577 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660598 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660618 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660636 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660685 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660720 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660751 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660792 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660820 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660860 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660892 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660918 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660944 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660979 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661001 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661030 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661064 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661090 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661112 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661156 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661181 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661204 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661226 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661248 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661269 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661329 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661352 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661381 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661408 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661440 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661464 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661507 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661554 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661577 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661601 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661629 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661665 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661688 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661711 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661799 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661830 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661861 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661892 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661915 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661937 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661964 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661985 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.662012 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.662038 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.662074 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.662098 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.662119 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.662157 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.662179 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.662201 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.662223 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.662255 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.662280 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.662309 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.662346 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.662367 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.662389 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.662412 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.662475 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.662500 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.662995 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.663091 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.663142 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.663171 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.663224 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.663251 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.663276 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.663301 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.663372 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.663403 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.663429 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.663455 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.663487 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.663520 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.663561 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.663586 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.663618 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.663643 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.663676 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.663699 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.663724 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.663747 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.663781 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.663817 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.663849 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.663873 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.663967 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.663995 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.664018 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.664043 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.664068 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.664093 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.664136 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.664163 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.664230 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.664257 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.664290 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.664322 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.664346 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.664370 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.664399 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.664423 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.664450 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.664476 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.664510 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.664554 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.664581 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.664607 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.664662 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.664697 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.664724 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.664755 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.664779 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.664864 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.664912 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.664939 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/8f28414c-12c1-4adb-be7b-6182310828eb-serviceca\") pod \"node-ca-tp5hk\" (UID: \"8f28414c-12c1-4adb-be7b-6182310828eb\") " pod="openshift-image-registry/node-ca-tp5hk" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.664976 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5swq\" (UniqueName: \"kubernetes.io/projected/8d21ac89-2ebd-49c3-9fe0-6c3f352d2257-kube-api-access-p5swq\") pod \"node-resolver-t4fq2\" (UID: \"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\") " pod="openshift-dns/node-resolver-t4fq2" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.665003 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.665044 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.665067 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzc59\" (UniqueName: \"kubernetes.io/projected/8f28414c-12c1-4adb-be7b-6182310828eb-kube-api-access-zzc59\") pod \"node-ca-tp5hk\" (UID: \"8f28414c-12c1-4adb-be7b-6182310828eb\") " pod="openshift-image-registry/node-ca-tp5hk" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.665093 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.665116 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8f28414c-12c1-4adb-be7b-6182310828eb-host\") pod \"node-ca-tp5hk\" (UID: \"8f28414c-12c1-4adb-be7b-6182310828eb\") " pod="openshift-image-registry/node-ca-tp5hk" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.665148 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.665182 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.665239 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.665267 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.665291 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8d21ac89-2ebd-49c3-9fe0-6c3f352d2257-hosts-file\") pod \"node-resolver-t4fq2\" (UID: \"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\") " pod="openshift-dns/node-resolver-t4fq2" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.665318 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.665343 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.665367 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.665405 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.665430 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.665553 4856 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.665578 4856 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.665595 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.665609 4856 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.665623 4856 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.665637 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.665652 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659764 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659848 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659864 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.659990 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660009 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660019 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660152 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660257 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660342 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660348 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660510 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660705 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660708 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660866 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.660910 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661021 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661187 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661182 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661391 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661408 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661505 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661611 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661676 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661706 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661800 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661904 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.661964 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.662032 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.662060 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.662108 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.662227 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.662483 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.662493 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.662817 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.662863 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.663025 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.663059 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.663094 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.663277 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.663363 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.663373 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.663378 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.663485 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.663624 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.663782 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.663802 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.664020 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.664059 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.664269 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.664421 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.664458 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.664490 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.664690 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.665899 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.666031 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.666547 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.666769 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.667034 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.667144 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.667143 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.667188 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.667223 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.667592 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.667865 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.668006 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.668069 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.668077 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.668090 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.668351 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.669133 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.669210 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.669329 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.669450 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.669511 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.669680 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.669617 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.670214 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.670480 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.671503 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.672726 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.672895 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.673059 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.673438 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.673919 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.674314 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.674440 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.674802 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.674947 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.679857 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.680266 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.680545 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.680815 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.681923 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.682591 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.682999 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.683235 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.683630 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.684065 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.685417 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.687768 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.688080 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.688893 4856 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Jan 26 16:58:54 crc kubenswrapper[4856]: E0126 16:58:54.689171 4856 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.689232 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.689261 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.689500 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.689910 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.693184 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.703786 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.724075 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.724345 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: E0126 16:58:54.724478 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:58:55.19097717 +0000 UTC m=+31.144231151 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.724679 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.725431 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.726795 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.729933 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.730007 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.730303 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.688826 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.730441 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.726477 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.730784 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.688885 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.730687 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.731035 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.731416 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.730884 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.732769 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.733115 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.733190 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.689094 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.733484 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: E0126 16:58:54.733500 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 16:58:55.233479199 +0000 UTC m=+31.186733180 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.733554 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.733801 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.733928 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.733920 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.734033 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.734051 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.734096 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.734284 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.734442 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.734504 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.734849 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.734971 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: E0126 16:58:54.735082 4856 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 16:58:54 crc kubenswrapper[4856]: E0126 16:58:54.735153 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 16:58:55.235131344 +0000 UTC m=+31.188385325 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.735454 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.735646 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.735663 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.735869 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.736110 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.736376 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.736608 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.736637 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.736996 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.737138 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.737243 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.737323 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.737477 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.737575 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.737805 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.738231 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.738309 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.738548 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.738900 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.739029 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.739466 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.739468 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.739546 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.739923 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.740003 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.747101 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.747384 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.747775 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.748102 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.748374 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.748651 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.748759 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.749085 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.749264 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.749273 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.749497 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: E0126 16:58:54.749646 4856 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 16:58:54 crc kubenswrapper[4856]: E0126 16:58:54.749707 4856 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.749719 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: E0126 16:58:54.749728 4856 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:58:54 crc kubenswrapper[4856]: E0126 16:58:54.750049 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-26 16:58:55.250014753 +0000 UTC m=+31.203268924 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.750155 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.750175 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.750194 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.750279 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.750290 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.704487 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.698382 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.750711 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.750777 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.755731 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.757567 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.759086 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.759450 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.759843 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:58:54 crc kubenswrapper[4856]: E0126 16:58:54.759965 4856 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 16:58:54 crc kubenswrapper[4856]: E0126 16:58:54.760023 4856 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 16:58:54 crc kubenswrapper[4856]: E0126 16:58:54.760049 4856 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:58:54 crc kubenswrapper[4856]: E0126 16:58:54.760213 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-26 16:58:55.260179322 +0000 UTC m=+31.213433463 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.764902 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.767557 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.767993 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzc59\" (UniqueName: \"kubernetes.io/projected/8f28414c-12c1-4adb-be7b-6182310828eb-kube-api-access-zzc59\") pod \"node-ca-tp5hk\" (UID: \"8f28414c-12c1-4adb-be7b-6182310828eb\") " pod="openshift-image-registry/node-ca-tp5hk" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768027 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8d21ac89-2ebd-49c3-9fe0-6c3f352d2257-hosts-file\") pod \"node-resolver-t4fq2\" (UID: \"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\") " pod="openshift-dns/node-resolver-t4fq2" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768046 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8f28414c-12c1-4adb-be7b-6182310828eb-host\") pod \"node-ca-tp5hk\" (UID: \"8f28414c-12c1-4adb-be7b-6182310828eb\") " pod="openshift-image-registry/node-ca-tp5hk" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768085 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768100 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768114 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/8f28414c-12c1-4adb-be7b-6182310828eb-serviceca\") pod \"node-ca-tp5hk\" (UID: \"8f28414c-12c1-4adb-be7b-6182310828eb\") " pod="openshift-image-registry/node-ca-tp5hk" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768128 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5swq\" (UniqueName: \"kubernetes.io/projected/8d21ac89-2ebd-49c3-9fe0-6c3f352d2257-kube-api-access-p5swq\") pod \"node-resolver-t4fq2\" (UID: \"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\") " pod="openshift-dns/node-resolver-t4fq2" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768173 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768187 4856 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768201 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768213 4856 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768224 4856 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768234 4856 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768245 4856 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768255 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768266 4856 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768276 4856 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768285 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768293 4856 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768302 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768310 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768320 4856 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768329 4856 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768337 4856 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768346 4856 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768354 4856 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768362 4856 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768371 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768383 4856 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768392 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768402 4856 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768410 4856 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768418 4856 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768426 4856 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768434 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768442 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768451 4856 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768459 4856 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768467 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768477 4856 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768485 4856 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768494 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768501 4856 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768509 4856 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768518 4856 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768541 4856 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768550 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768559 4856 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768567 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768576 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768585 4856 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768594 4856 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768602 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768611 4856 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768621 4856 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768629 4856 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768637 4856 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768645 4856 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768654 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768662 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768672 4856 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768680 4856 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768688 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768696 4856 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768704 4856 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768715 4856 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768724 4856 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768722 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768734 4856 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768771 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768790 4856 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768804 4856 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768818 4856 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768830 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768842 4856 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768853 4856 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768865 4856 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768876 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768888 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768900 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768925 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768937 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768950 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768965 4856 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768983 4856 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.768997 4856 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.769487 4856 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.769510 4856 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.769522 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.769581 4856 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.769592 4856 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.769601 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/8d21ac89-2ebd-49c3-9fe0-6c3f352d2257-hosts-file\") pod \"node-resolver-t4fq2\" (UID: \"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\") " pod="openshift-dns/node-resolver-t4fq2" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.769605 4856 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.769637 4856 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.769643 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.769685 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.769746 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8f28414c-12c1-4adb-be7b-6182310828eb-host\") pod \"node-ca-tp5hk\" (UID: \"8f28414c-12c1-4adb-be7b-6182310828eb\") " pod="openshift-image-registry/node-ca-tp5hk" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.769649 4856 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.769799 4856 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.769812 4856 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.769824 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.769836 4856 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.769849 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.769861 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.769873 4856 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.769883 4856 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.769895 4856 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.769906 4856 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.769923 4856 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.769933 4856 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.769944 4856 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.769954 4856 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.769964 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.769974 4856 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.769983 4856 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.769992 4856 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770002 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770014 4856 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770026 4856 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770038 4856 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770049 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770060 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770071 4856 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770081 4856 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770091 4856 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770102 4856 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770112 4856 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770122 4856 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770133 4856 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770146 4856 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770157 4856 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770180 4856 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770192 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770204 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770216 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770228 4856 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770245 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770257 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770267 4856 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770279 4856 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770289 4856 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770300 4856 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770313 4856 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770337 4856 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770349 4856 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770361 4856 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770373 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770385 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770396 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770409 4856 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770420 4856 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770430 4856 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770442 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770452 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770462 4856 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770474 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770484 4856 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770494 4856 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770503 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770513 4856 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770538 4856 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770550 4856 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770559 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770568 4856 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770576 4856 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770583 4856 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770593 4856 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770602 4856 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770610 4856 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770619 4856 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770628 4856 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770640 4856 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770647 4856 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770656 4856 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770664 4856 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770672 4856 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770681 4856 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770691 4856 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770700 4856 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770709 4856 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770718 4856 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770726 4856 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770735 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770744 4856 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770753 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770760 4856 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770768 4856 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770776 4856 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770785 4856 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770793 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770801 4856 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770809 4856 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770817 4856 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770825 4856 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770835 4856 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770843 4856 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.770851 4856 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.773225 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.773566 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/8f28414c-12c1-4adb-be7b-6182310828eb-serviceca\") pod \"node-ca-tp5hk\" (UID: \"8f28414c-12c1-4adb-be7b-6182310828eb\") " pod="openshift-image-registry/node-ca-tp5hk" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.782486 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.792161 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.796175 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzc59\" (UniqueName: \"kubernetes.io/projected/8f28414c-12c1-4adb-be7b-6182310828eb-kube-api-access-zzc59\") pod \"node-ca-tp5hk\" (UID: \"8f28414c-12c1-4adb-be7b-6182310828eb\") " pod="openshift-image-registry/node-ca-tp5hk" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.796732 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.797827 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5swq\" (UniqueName: \"kubernetes.io/projected/8d21ac89-2ebd-49c3-9fe0-6c3f352d2257-kube-api-access-p5swq\") pod \"node-resolver-t4fq2\" (UID: \"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\") " pod="openshift-dns/node-resolver-t4fq2" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.812432 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.823562 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.837066 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.847323 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.852596 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-tp5hk" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.863097 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.863774 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:58:54 crc kubenswrapper[4856]: W0126 16:58:54.867994 4856 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f28414c_12c1_4adb_be7b_6182310828eb.slice/crio-a37e0c99cc001b0c3bf3066fb8d729ead73f8dac4bab47ada43a1f8f0f83aeb9 WatchSource:0}: Error finding container a37e0c99cc001b0c3bf3066fb8d729ead73f8dac4bab47ada43a1f8f0f83aeb9: Status 404 returned error can't find the container with id a37e0c99cc001b0c3bf3066fb8d729ead73f8dac4bab47ada43a1f8f0f83aeb9 Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.871620 4856 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.871646 4856 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.877232 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.879421 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.895223 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 26 16:58:54 crc kubenswrapper[4856]: I0126 16:58:54.895230 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:54.901879 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-t4fq2" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:54.931772 4856 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Jan 26 16:58:55 crc kubenswrapper[4856]: W0126 16:58:54.931967 4856 reflector.go:484] object-"openshift-network-operator"/"iptables-alerter-script": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-operator"/"iptables-alerter-script": Unexpected watch close - watch lasted less than a second and no items received Jan 26 16:58:55 crc kubenswrapper[4856]: W0126 16:58:54.932001 4856 reflector.go:484] object-"openshift-network-node-identity"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 26 16:58:55 crc kubenswrapper[4856]: W0126 16:58:54.932022 4856 reflector.go:484] object-"openshift-network-node-identity"/"ovnkube-identity-cm": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"ovnkube-identity-cm": Unexpected watch close - watch lasted less than a second and no items received Jan 26 16:58:55 crc kubenswrapper[4856]: W0126 16:58:54.932069 4856 reflector.go:484] object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": watch of *v1.Secret ended with: very short watch: object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": Unexpected watch close - watch lasted less than a second and no items received Jan 26 16:58:55 crc kubenswrapper[4856]: E0126 16:58:54.932123 4856 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-node-identity/events\": read tcp 38.102.83.241:34646->38.102.83.241:6443: use of closed network connection" event="&Event{ObjectMeta:{network-node-identity-vrzqb.188e566e339fe4da openshift-network-node-identity 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-network-node-identity,Name:network-node-identity-vrzqb,UID:ef543e1b-8068-4ea3-b32a-61027b32e95d,APIVersion:v1,ResourceVersion:25324,FieldPath:spec.containers{webhook},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-26 16:58:54.915691738 +0000 UTC m=+30.868945709,LastTimestamp:2026-01-26 16:58:54.915691738 +0000 UTC m=+30.868945709,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 26 16:58:55 crc kubenswrapper[4856]: W0126 16:58:54.932260 4856 reflector.go:484] object-"openshift-dns"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-dns"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 26 16:58:55 crc kubenswrapper[4856]: W0126 16:58:54.932281 4856 reflector.go:484] object-"openshift-network-operator"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-operator"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 26 16:58:55 crc kubenswrapper[4856]: W0126 16:58:54.932303 4856 reflector.go:484] object-"openshift-network-operator"/"metrics-tls": watch of *v1.Secret ended with: very short watch: object-"openshift-network-operator"/"metrics-tls": Unexpected watch close - watch lasted less than a second and no items received Jan 26 16:58:55 crc kubenswrapper[4856]: W0126 16:58:54.932323 4856 reflector.go:484] object-"openshift-dns"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-dns"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 26 16:58:55 crc kubenswrapper[4856]: W0126 16:58:54.932348 4856 reflector.go:484] k8s.io/client-go/informers/factory.go:160: watch of *v1.CSIDriver ended with: very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received Jan 26 16:58:55 crc kubenswrapper[4856]: W0126 16:58:54.932394 4856 reflector.go:484] object-"openshift-image-registry"/"node-ca-dockercfg-4777p": watch of *v1.Secret ended with: very short watch: object-"openshift-image-registry"/"node-ca-dockercfg-4777p": Unexpected watch close - watch lasted less than a second and no items received Jan 26 16:58:55 crc kubenswrapper[4856]: W0126 16:58:54.932416 4856 reflector.go:484] object-"openshift-network-node-identity"/"env-overrides": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"env-overrides": Unexpected watch close - watch lasted less than a second and no items received Jan 26 16:58:55 crc kubenswrapper[4856]: W0126 16:58:54.933100 4856 reflector.go:484] pkg/kubelet/config/apiserver.go:66: watch of *v1.Pod ended with: very short watch: pkg/kubelet/config/apiserver.go:66: Unexpected watch close - watch lasted less than a second and no items received Jan 26 16:58:55 crc kubenswrapper[4856]: W0126 16:58:54.933126 4856 reflector.go:484] object-"openshift-network-node-identity"/"network-node-identity-cert": watch of *v1.Secret ended with: very short watch: object-"openshift-network-node-identity"/"network-node-identity-cert": Unexpected watch close - watch lasted less than a second and no items received Jan 26 16:58:55 crc kubenswrapper[4856]: W0126 16:58:54.933148 4856 reflector.go:484] object-"openshift-image-registry"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-image-registry"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 26 16:58:55 crc kubenswrapper[4856]: W0126 16:58:54.933183 4856 reflector.go:484] object-"openshift-network-operator"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-operator"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 26 16:58:55 crc kubenswrapper[4856]: W0126 16:58:54.933202 4856 reflector.go:484] object-"openshift-image-registry"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-image-registry"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 26 16:58:55 crc kubenswrapper[4856]: W0126 16:58:54.933248 4856 reflector.go:484] object-"openshift-image-registry"/"image-registry-certificates": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-image-registry"/"image-registry-certificates": Unexpected watch close - watch lasted less than a second and no items received Jan 26 16:58:55 crc kubenswrapper[4856]: W0126 16:58:54.933269 4856 reflector.go:484] object-"openshift-network-node-identity"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Jan 26 16:58:55 crc kubenswrapper[4856]: W0126 16:58:55.068116 4856 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-8e96a123a536cdbe669eb3f2682c4e4aa9f0a6ed84a372d93b59d4ff124bfbdd WatchSource:0}: Error finding container 8e96a123a536cdbe669eb3f2682c4e4aa9f0a6ed84a372d93b59d4ff124bfbdd: Status 404 returned error can't find the container with id 8e96a123a536cdbe669eb3f2682c4e4aa9f0a6ed84a372d93b59d4ff124bfbdd Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.274131 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.274253 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.274304 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:58:55 crc kubenswrapper[4856]: E0126 16:58:55.274391 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:58:56.274356378 +0000 UTC m=+32.227610359 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:58:55 crc kubenswrapper[4856]: E0126 16:58:55.274482 4856 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 16:58:55 crc kubenswrapper[4856]: E0126 16:58:55.274541 4856 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 16:58:55 crc kubenswrapper[4856]: E0126 16:58:55.274556 4856 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.274516 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:58:55 crc kubenswrapper[4856]: E0126 16:58:55.274624 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-26 16:58:56.274606484 +0000 UTC m=+32.227860465 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.274650 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:58:55 crc kubenswrapper[4856]: E0126 16:58:55.274677 4856 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 16:58:55 crc kubenswrapper[4856]: E0126 16:58:55.274710 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 16:58:56.274701687 +0000 UTC m=+32.227955668 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 16:58:55 crc kubenswrapper[4856]: E0126 16:58:55.274775 4856 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 16:58:55 crc kubenswrapper[4856]: E0126 16:58:55.274797 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 16:58:56.27479159 +0000 UTC m=+32.228045571 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 16:58:55 crc kubenswrapper[4856]: E0126 16:58:55.274881 4856 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 16:58:55 crc kubenswrapper[4856]: E0126 16:58:55.274903 4856 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 16:58:55 crc kubenswrapper[4856]: E0126 16:58:55.274922 4856 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:58:55 crc kubenswrapper[4856]: E0126 16:58:55.274976 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-26 16:58:56.274965764 +0000 UTC m=+32.228219925 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.412723 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.413824 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.415807 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.416941 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.419601 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.420630 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.421831 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.423566 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.424451 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.425945 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.426905 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.428218 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.430465 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.431576 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.432945 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.433773 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.434435 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.436101 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.436749 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.437604 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.438598 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.439845 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.440727 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.443033 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.443686 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.445217 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.445841 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.447768 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.448803 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.449400 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.450981 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.451670 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.452835 4856 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.452826 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.453152 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.455295 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.457391 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.457969 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 17:38:09.462757487 +0000 UTC Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.458783 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.460887 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.462128 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.463242 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.464835 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.465897 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.466448 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.467689 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.468894 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.469595 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.470430 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.470957 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.472065 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.472795 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.473704 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.464855 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.474214 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.474707 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.475646 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.476209 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.477075 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.498452 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.564402 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.580648 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.589908 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.870238 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-t4fq2" event={"ID":"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257","Type":"ContainerStarted","Data":"627cd6b39fdedd967fffcfc0755439277b2d73016fba1ddc7f5b9d5deba43b8f"} Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.870284 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-t4fq2" event={"ID":"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257","Type":"ContainerStarted","Data":"0daeec337d442d6b72d206a5437a7229a7184f72a23ece576a2bf30bb2aee119"} Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.871271 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"8e96a123a536cdbe669eb3f2682c4e4aa9f0a6ed84a372d93b59d4ff124bfbdd"} Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.873743 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"191ad5d41024c88c1a4fbc30f307dd6340fd55b93b00d22f62209d0e82be286f"} Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.873809 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ccf2a4d8be409a46ffe5702797e56a80023a72a19fb7dc5c49e5f4984cedc600"} Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.873841 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"47335579ac42d26a6a37c7e5cae4b0819f8404d34215632c9ea2b86bc72395da"} Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.875845 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.877215 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"3c09748c0963bdb50c9552f249d1135ea53e68c52babd788dad050951cb849cf"} Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.877289 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"1e97d53e095cbf2a8b3483f26f64d24c54d4eca3ae638425541067dd3c3e6c08"} Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.878751 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-tp5hk" event={"ID":"8f28414c-12c1-4adb-be7b-6182310828eb","Type":"ContainerStarted","Data":"0af4930ff9cebd62106545fb3da2dfc93f7b591426fe4c85aa2e637b60c935f7"} Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.878796 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-tp5hk" event={"ID":"8f28414c-12c1-4adb-be7b-6182310828eb","Type":"ContainerStarted","Data":"a37e0c99cc001b0c3bf3066fb8d729ead73f8dac4bab47ada43a1f8f0f83aeb9"} Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.883731 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.884172 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Jan 26 16:58:55 crc kubenswrapper[4856]: I0126 16:58:55.906972 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.018598 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.021580 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.030014 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.040461 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.041885 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.053004 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.063132 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://627cd6b39fdedd967fffcfc0755439277b2d73016fba1ddc7f5b9d5deba43b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.074835 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.084737 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.099127 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.119624 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.130377 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af4930ff9cebd62106545fb3da2dfc93f7b591426fe4c85aa2e637b60c935f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.143422 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c09748c0963bdb50c9552f249d1135ea53e68c52babd788dad050951cb849cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.145691 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.149166 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.156094 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.164032 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.167717 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191ad5d41024c88c1a4fbc30f307dd6340fd55b93b00d22f62209d0e82be286f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2a4d8be409a46ffe5702797e56a80023a72a19fb7dc5c49e5f4984cedc600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.179940 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.191635 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.202637 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://627cd6b39fdedd967fffcfc0755439277b2d73016fba1ddc7f5b9d5deba43b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.271648 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.297117 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:58:56 crc kubenswrapper[4856]: E0126 16:58:56.297400 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:58:58.29736032 +0000 UTC m=+34.250614321 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.297472 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.297565 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.297605 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.297645 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:58:56 crc kubenswrapper[4856]: E0126 16:58:56.297785 4856 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 16:58:56 crc kubenswrapper[4856]: E0126 16:58:56.297839 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 16:58:58.297828303 +0000 UTC m=+34.251082284 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 16:58:56 crc kubenswrapper[4856]: E0126 16:58:56.297847 4856 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 16:58:56 crc kubenswrapper[4856]: E0126 16:58:56.297867 4856 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 16:58:56 crc kubenswrapper[4856]: E0126 16:58:56.297880 4856 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:58:56 crc kubenswrapper[4856]: E0126 16:58:56.297917 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-26 16:58:58.297903915 +0000 UTC m=+34.251157896 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:58:56 crc kubenswrapper[4856]: E0126 16:58:56.298015 4856 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 16:58:56 crc kubenswrapper[4856]: E0126 16:58:56.298075 4856 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 16:58:56 crc kubenswrapper[4856]: E0126 16:58:56.298097 4856 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:58:56 crc kubenswrapper[4856]: E0126 16:58:56.298175 4856 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 16:58:56 crc kubenswrapper[4856]: E0126 16:58:56.298193 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-26 16:58:58.298164813 +0000 UTC m=+34.251418814 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:58:56 crc kubenswrapper[4856]: E0126 16:58:56.298366 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 16:58:58.298345127 +0000 UTC m=+34.251599108 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.367989 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.394938 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.394974 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.395008 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:58:56 crc kubenswrapper[4856]: E0126 16:58:56.395059 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:58:56 crc kubenswrapper[4856]: E0126 16:58:56.395141 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:58:56 crc kubenswrapper[4856]: E0126 16:58:56.395249 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.396683 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.419888 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.429918 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.458802 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 01:01:21.939772445 +0000 UTC Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.466305 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.472889 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.595478 4856 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.596341 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-v2l7v","openshift-ovn-kubernetes/ovnkube-node-pxh94","openshift-machine-config-operator/machine-config-daemon-xm9cq","openshift-multus/multus-rq622"] Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.597376 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.599165 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.601029 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.601131 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.601177 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.601349 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.602250 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.613356 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.613430 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.613964 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.614629 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.614707 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.614870 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.616044 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.616206 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.616369 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.616488 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.616541 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.616582 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.616824 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.617031 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.617196 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.617362 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.624833 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:56Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.637807 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af4930ff9cebd62106545fb3da2dfc93f7b591426fe4c85aa2e637b60c935f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:56Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.654829 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c09748c0963bdb50c9552f249d1135ea53e68c52babd788dad050951cb849cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:56Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.670358 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:56Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.683617 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:56Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.695041 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://627cd6b39fdedd967fffcfc0755439277b2d73016fba1ddc7f5b9d5deba43b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:56Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.700488 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-systemd-units\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.700569 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7a742e7b-c420-46e3-9e96-e9c744af6124-etc-kubernetes\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.700590 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-ovnkube-config\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.700612 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zm9x6\" (UniqueName: \"kubernetes.io/projected/ad7b59f9-beb7-49d6-a2d1-e29133e46854-kube-api-access-zm9x6\") pod \"multus-additional-cni-plugins-v2l7v\" (UID: \"ad7b59f9-beb7-49d6-a2d1-e29133e46854\") " pod="openshift-multus/multus-additional-cni-plugins-v2l7v" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.700632 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/63c75ede-5170-4db0-811b-5217ef8d72b3-proxy-tls\") pod \"machine-config-daemon-xm9cq\" (UID: \"63c75ede-5170-4db0-811b-5217ef8d72b3\") " pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.700649 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-run-openvswitch\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.700665 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/7a742e7b-c420-46e3-9e96-e9c744af6124-multus-daemon-config\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.700772 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ad7b59f9-beb7-49d6-a2d1-e29133e46854-cnibin\") pod \"multus-additional-cni-plugins-v2l7v\" (UID: \"ad7b59f9-beb7-49d6-a2d1-e29133e46854\") " pod="openshift-multus/multus-additional-cni-plugins-v2l7v" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.700842 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-run-systemd\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.700872 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7a742e7b-c420-46e3-9e96-e9c744af6124-multus-cni-dir\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.700889 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7a742e7b-c420-46e3-9e96-e9c744af6124-os-release\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.700931 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-host-kubelet\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.700951 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-node-log\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.700968 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/7a742e7b-c420-46e3-9e96-e9c744af6124-hostroot\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.700985 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ad7b59f9-beb7-49d6-a2d1-e29133e46854-cni-binary-copy\") pod \"multus-additional-cni-plugins-v2l7v\" (UID: \"ad7b59f9-beb7-49d6-a2d1-e29133e46854\") " pod="openshift-multus/multus-additional-cni-plugins-v2l7v" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.701007 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-host-cni-bin\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.701025 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kdbz\" (UniqueName: \"kubernetes.io/projected/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-kube-api-access-9kdbz\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.701044 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7a742e7b-c420-46e3-9e96-e9c744af6124-cnibin\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.701058 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ad7b59f9-beb7-49d6-a2d1-e29133e46854-system-cni-dir\") pod \"multus-additional-cni-plugins-v2l7v\" (UID: \"ad7b59f9-beb7-49d6-a2d1-e29133e46854\") " pod="openshift-multus/multus-additional-cni-plugins-v2l7v" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.701074 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/ad7b59f9-beb7-49d6-a2d1-e29133e46854-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-v2l7v\" (UID: \"ad7b59f9-beb7-49d6-a2d1-e29133e46854\") " pod="openshift-multus/multus-additional-cni-plugins-v2l7v" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.701119 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7a742e7b-c420-46e3-9e96-e9c744af6124-system-cni-dir\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.701146 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ad7b59f9-beb7-49d6-a2d1-e29133e46854-tuning-conf-dir\") pod \"multus-additional-cni-plugins-v2l7v\" (UID: \"ad7b59f9-beb7-49d6-a2d1-e29133e46854\") " pod="openshift-multus/multus-additional-cni-plugins-v2l7v" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.701164 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/63c75ede-5170-4db0-811b-5217ef8d72b3-rootfs\") pod \"machine-config-daemon-xm9cq\" (UID: \"63c75ede-5170-4db0-811b-5217ef8d72b3\") " pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.701185 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96lw2\" (UniqueName: \"kubernetes.io/projected/63c75ede-5170-4db0-811b-5217ef8d72b3-kube-api-access-96lw2\") pod \"machine-config-daemon-xm9cq\" (UID: \"63c75ede-5170-4db0-811b-5217ef8d72b3\") " pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.701215 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/7a742e7b-c420-46e3-9e96-e9c744af6124-host-run-k8s-cni-cncf-io\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.701246 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-etc-openvswitch\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.701302 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-host-run-ovn-kubernetes\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.701363 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7a742e7b-c420-46e3-9e96-e9c744af6124-cni-binary-copy\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.701378 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7a742e7b-c420-46e3-9e96-e9c744af6124-host-var-lib-cni-bin\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.701426 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ad7b59f9-beb7-49d6-a2d1-e29133e46854-os-release\") pod \"multus-additional-cni-plugins-v2l7v\" (UID: \"ad7b59f9-beb7-49d6-a2d1-e29133e46854\") " pod="openshift-multus/multus-additional-cni-plugins-v2l7v" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.701462 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.701494 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/7a742e7b-c420-46e3-9e96-e9c744af6124-host-run-multus-certs\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.701544 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-run-ovn\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.701572 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-log-socket\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.701600 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-ovnkube-script-lib\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.701632 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/7a742e7b-c420-46e3-9e96-e9c744af6124-multus-socket-dir-parent\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.701649 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7a742e7b-c420-46e3-9e96-e9c744af6124-host-run-netns\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.701669 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/7a742e7b-c420-46e3-9e96-e9c744af6124-host-var-lib-kubelet\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.701686 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7a742e7b-c420-46e3-9e96-e9c744af6124-multus-conf-dir\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.701701 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/63c75ede-5170-4db0-811b-5217ef8d72b3-mcd-auth-proxy-config\") pod \"machine-config-daemon-xm9cq\" (UID: \"63c75ede-5170-4db0-811b-5217ef8d72b3\") " pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.701780 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-host-slash\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.701805 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-host-cni-netd\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.701828 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-ovn-node-metrics-cert\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.701851 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/7a742e7b-c420-46e3-9e96-e9c744af6124-host-var-lib-cni-multus\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.701881 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8plh8\" (UniqueName: \"kubernetes.io/projected/7a742e7b-c420-46e3-9e96-e9c744af6124-kube-api-access-8plh8\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.701926 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-host-run-netns\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.701957 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-var-lib-openvswitch\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.701991 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-env-overrides\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.707997 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191ad5d41024c88c1a4fbc30f307dd6340fd55b93b00d22f62209d0e82be286f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2a4d8be409a46ffe5702797e56a80023a72a19fb7dc5c49e5f4984cedc600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:56Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.719987 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:56Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.734439 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad7b59f9-beb7-49d6-a2d1-e29133e46854\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v2l7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:56Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.747574 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191ad5d41024c88c1a4fbc30f307dd6340fd55b93b00d22f62209d0e82be286f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2a4d8be409a46ffe5702797e56a80023a72a19fb7dc5c49e5f4984cedc600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:56Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.760933 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:56Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.773800 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:56Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.783472 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://627cd6b39fdedd967fffcfc0755439277b2d73016fba1ddc7f5b9d5deba43b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:56Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.796061 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rq622" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a742e7b-c420-46e3-9e96-e9c744af6124\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8plh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rq622\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:56Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.803547 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.803600 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/7a742e7b-c420-46e3-9e96-e9c744af6124-host-run-multus-certs\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.803623 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-run-ovn\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.803643 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-log-socket\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.803675 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-ovnkube-script-lib\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.803695 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/7a742e7b-c420-46e3-9e96-e9c744af6124-multus-socket-dir-parent\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.803714 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7a742e7b-c420-46e3-9e96-e9c744af6124-host-run-netns\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.803735 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/7a742e7b-c420-46e3-9e96-e9c744af6124-host-var-lib-kubelet\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.803754 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7a742e7b-c420-46e3-9e96-e9c744af6124-multus-conf-dir\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.803747 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.803799 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7a742e7b-c420-46e3-9e96-e9c744af6124-host-run-netns\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.803795 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/7a742e7b-c420-46e3-9e96-e9c744af6124-host-run-multus-certs\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.803777 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/63c75ede-5170-4db0-811b-5217ef8d72b3-mcd-auth-proxy-config\") pod \"machine-config-daemon-xm9cq\" (UID: \"63c75ede-5170-4db0-811b-5217ef8d72b3\") " pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.803758 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-run-ovn\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.803877 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-log-socket\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.803891 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7a742e7b-c420-46e3-9e96-e9c744af6124-multus-conf-dir\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.803886 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/7a742e7b-c420-46e3-9e96-e9c744af6124-host-var-lib-kubelet\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.803956 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/7a742e7b-c420-46e3-9e96-e9c744af6124-multus-socket-dir-parent\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804050 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-host-slash\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804089 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-host-cni-netd\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804116 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-ovn-node-metrics-cert\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804132 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-host-slash\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804156 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-host-cni-netd\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804166 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/7a742e7b-c420-46e3-9e96-e9c744af6124-host-var-lib-cni-multus\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804139 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/7a742e7b-c420-46e3-9e96-e9c744af6124-host-var-lib-cni-multus\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804209 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8plh8\" (UniqueName: \"kubernetes.io/projected/7a742e7b-c420-46e3-9e96-e9c744af6124-kube-api-access-8plh8\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804259 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-host-run-netns\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804281 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-var-lib-openvswitch\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804301 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-env-overrides\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804329 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-systemd-units\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804331 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-var-lib-openvswitch\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804356 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7a742e7b-c420-46e3-9e96-e9c744af6124-etc-kubernetes\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804385 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zm9x6\" (UniqueName: \"kubernetes.io/projected/ad7b59f9-beb7-49d6-a2d1-e29133e46854-kube-api-access-zm9x6\") pod \"multus-additional-cni-plugins-v2l7v\" (UID: \"ad7b59f9-beb7-49d6-a2d1-e29133e46854\") " pod="openshift-multus/multus-additional-cni-plugins-v2l7v" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804422 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/63c75ede-5170-4db0-811b-5217ef8d72b3-proxy-tls\") pod \"machine-config-daemon-xm9cq\" (UID: \"63c75ede-5170-4db0-811b-5217ef8d72b3\") " pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804421 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-systemd-units\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804440 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-ovnkube-config\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804475 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-run-openvswitch\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804503 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/7a742e7b-c420-46e3-9e96-e9c744af6124-multus-daemon-config\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804502 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7a742e7b-c420-46e3-9e96-e9c744af6124-etc-kubernetes\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804573 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ad7b59f9-beb7-49d6-a2d1-e29133e46854-cnibin\") pod \"multus-additional-cni-plugins-v2l7v\" (UID: \"ad7b59f9-beb7-49d6-a2d1-e29133e46854\") " pod="openshift-multus/multus-additional-cni-plugins-v2l7v" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804546 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ad7b59f9-beb7-49d6-a2d1-e29133e46854-cnibin\") pod \"multus-additional-cni-plugins-v2l7v\" (UID: \"ad7b59f9-beb7-49d6-a2d1-e29133e46854\") " pod="openshift-multus/multus-additional-cni-plugins-v2l7v" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804628 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-run-systemd\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804653 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7a742e7b-c420-46e3-9e96-e9c744af6124-multus-cni-dir\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804672 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7a742e7b-c420-46e3-9e96-e9c744af6124-os-release\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804716 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-host-kubelet\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804733 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-node-log\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804748 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/7a742e7b-c420-46e3-9e96-e9c744af6124-hostroot\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804773 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-host-cni-bin\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804790 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kdbz\" (UniqueName: \"kubernetes.io/projected/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-kube-api-access-9kdbz\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804808 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7a742e7b-c420-46e3-9e96-e9c744af6124-cnibin\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804826 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ad7b59f9-beb7-49d6-a2d1-e29133e46854-system-cni-dir\") pod \"multus-additional-cni-plugins-v2l7v\" (UID: \"ad7b59f9-beb7-49d6-a2d1-e29133e46854\") " pod="openshift-multus/multus-additional-cni-plugins-v2l7v" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804842 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ad7b59f9-beb7-49d6-a2d1-e29133e46854-cni-binary-copy\") pod \"multus-additional-cni-plugins-v2l7v\" (UID: \"ad7b59f9-beb7-49d6-a2d1-e29133e46854\") " pod="openshift-multus/multus-additional-cni-plugins-v2l7v" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804862 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/ad7b59f9-beb7-49d6-a2d1-e29133e46854-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-v2l7v\" (UID: \"ad7b59f9-beb7-49d6-a2d1-e29133e46854\") " pod="openshift-multus/multus-additional-cni-plugins-v2l7v" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804874 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/63c75ede-5170-4db0-811b-5217ef8d72b3-mcd-auth-proxy-config\") pod \"machine-config-daemon-xm9cq\" (UID: \"63c75ede-5170-4db0-811b-5217ef8d72b3\") " pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804884 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7a742e7b-c420-46e3-9e96-e9c744af6124-system-cni-dir\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804894 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-ovnkube-script-lib\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804908 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-env-overrides\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804877 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-run-openvswitch\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804905 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ad7b59f9-beb7-49d6-a2d1-e29133e46854-tuning-conf-dir\") pod \"multus-additional-cni-plugins-v2l7v\" (UID: \"ad7b59f9-beb7-49d6-a2d1-e29133e46854\") " pod="openshift-multus/multus-additional-cni-plugins-v2l7v" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804962 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-host-cni-bin\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804963 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ad7b59f9-beb7-49d6-a2d1-e29133e46854-system-cni-dir\") pod \"multus-additional-cni-plugins-v2l7v\" (UID: \"ad7b59f9-beb7-49d6-a2d1-e29133e46854\") " pod="openshift-multus/multus-additional-cni-plugins-v2l7v" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.804983 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/63c75ede-5170-4db0-811b-5217ef8d72b3-rootfs\") pod \"machine-config-daemon-xm9cq\" (UID: \"63c75ede-5170-4db0-811b-5217ef8d72b3\") " pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.805012 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/63c75ede-5170-4db0-811b-5217ef8d72b3-rootfs\") pod \"machine-config-daemon-xm9cq\" (UID: \"63c75ede-5170-4db0-811b-5217ef8d72b3\") " pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.805029 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96lw2\" (UniqueName: \"kubernetes.io/projected/63c75ede-5170-4db0-811b-5217ef8d72b3-kube-api-access-96lw2\") pod \"machine-config-daemon-xm9cq\" (UID: \"63c75ede-5170-4db0-811b-5217ef8d72b3\") " pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.805060 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/7a742e7b-c420-46e3-9e96-e9c744af6124-host-run-k8s-cni-cncf-io\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.805069 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7a742e7b-c420-46e3-9e96-e9c744af6124-cnibin\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.805104 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-host-kubelet\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.805105 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-etc-openvswitch\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.805131 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-run-systemd\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.805137 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-host-run-ovn-kubernetes\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.805157 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/7a742e7b-c420-46e3-9e96-e9c744af6124-hostroot\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.805168 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7a742e7b-c420-46e3-9e96-e9c744af6124-cni-binary-copy\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.805214 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7a742e7b-c420-46e3-9e96-e9c744af6124-host-var-lib-cni-bin\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.805243 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ad7b59f9-beb7-49d6-a2d1-e29133e46854-os-release\") pod \"multus-additional-cni-plugins-v2l7v\" (UID: \"ad7b59f9-beb7-49d6-a2d1-e29133e46854\") " pod="openshift-multus/multus-additional-cni-plugins-v2l7v" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.805332 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7a742e7b-c420-46e3-9e96-e9c744af6124-multus-cni-dir\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.805600 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/7a742e7b-c420-46e3-9e96-e9c744af6124-multus-daemon-config\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.805623 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-ovnkube-config\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.805623 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ad7b59f9-beb7-49d6-a2d1-e29133e46854-os-release\") pod \"multus-additional-cni-plugins-v2l7v\" (UID: \"ad7b59f9-beb7-49d6-a2d1-e29133e46854\") " pod="openshift-multus/multus-additional-cni-plugins-v2l7v" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.805630 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7a742e7b-c420-46e3-9e96-e9c744af6124-os-release\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.805651 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/7a742e7b-c420-46e3-9e96-e9c744af6124-host-run-k8s-cni-cncf-io\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.805661 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-node-log\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.805702 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7a742e7b-c420-46e3-9e96-e9c744af6124-host-var-lib-cni-bin\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.805746 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7a742e7b-c420-46e3-9e96-e9c744af6124-system-cni-dir\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.805769 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-etc-openvswitch\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.805781 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7a742e7b-c420-46e3-9e96-e9c744af6124-cni-binary-copy\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.805983 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ad7b59f9-beb7-49d6-a2d1-e29133e46854-tuning-conf-dir\") pod \"multus-additional-cni-plugins-v2l7v\" (UID: \"ad7b59f9-beb7-49d6-a2d1-e29133e46854\") " pod="openshift-multus/multus-additional-cni-plugins-v2l7v" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.805794 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-host-run-ovn-kubernetes\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.806297 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/ad7b59f9-beb7-49d6-a2d1-e29133e46854-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-v2l7v\" (UID: \"ad7b59f9-beb7-49d6-a2d1-e29133e46854\") " pod="openshift-multus/multus-additional-cni-plugins-v2l7v" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.806401 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ad7b59f9-beb7-49d6-a2d1-e29133e46854-cni-binary-copy\") pod \"multus-additional-cni-plugins-v2l7v\" (UID: \"ad7b59f9-beb7-49d6-a2d1-e29133e46854\") " pod="openshift-multus/multus-additional-cni-plugins-v2l7v" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.806472 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-host-run-netns\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.807648 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:56Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.812259 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-ovn-node-metrics-cert\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.815129 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/63c75ede-5170-4db0-811b-5217ef8d72b3-proxy-tls\") pod \"machine-config-daemon-xm9cq\" (UID: \"63c75ede-5170-4db0-811b-5217ef8d72b3\") " pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.821547 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zm9x6\" (UniqueName: \"kubernetes.io/projected/ad7b59f9-beb7-49d6-a2d1-e29133e46854-kube-api-access-zm9x6\") pod \"multus-additional-cni-plugins-v2l7v\" (UID: \"ad7b59f9-beb7-49d6-a2d1-e29133e46854\") " pod="openshift-multus/multus-additional-cni-plugins-v2l7v" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.822432 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96lw2\" (UniqueName: \"kubernetes.io/projected/63c75ede-5170-4db0-811b-5217ef8d72b3-kube-api-access-96lw2\") pod \"machine-config-daemon-xm9cq\" (UID: \"63c75ede-5170-4db0-811b-5217ef8d72b3\") " pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.826385 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kdbz\" (UniqueName: \"kubernetes.io/projected/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-kube-api-access-9kdbz\") pod \"ovnkube-node-pxh94\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.827042 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8plh8\" (UniqueName: \"kubernetes.io/projected/7a742e7b-c420-46e3-9e96-e9c744af6124-kube-api-access-8plh8\") pod \"multus-rq622\" (UID: \"7a742e7b-c420-46e3-9e96-e9c744af6124\") " pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.827057 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad7b59f9-beb7-49d6-a2d1-e29133e46854\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v2l7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:56Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.851799 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pxh94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:56Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.863697 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63c75ede-5170-4db0-811b-5217ef8d72b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xm9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:56Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.873006 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af4930ff9cebd62106545fb3da2dfc93f7b591426fe4c85aa2e637b60c935f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:56Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.893673 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c09748c0963bdb50c9552f249d1135ea53e68c52babd788dad050951cb849cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:56Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.905434 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:56Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.922720 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.929847 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.937861 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" Jan 26 16:58:56 crc kubenswrapper[4856]: I0126 16:58:56.944856 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-rq622" Jan 26 16:58:56 crc kubenswrapper[4856]: W0126 16:58:56.964612 4856 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod63c75ede_5170_4db0_811b_5217ef8d72b3.slice/crio-a4deba88fd2726a90167401cf17b82783a9a1e76e5ef43e68433893ec3aaa466 WatchSource:0}: Error finding container a4deba88fd2726a90167401cf17b82783a9a1e76e5ef43e68433893ec3aaa466: Status 404 returned error can't find the container with id a4deba88fd2726a90167401cf17b82783a9a1e76e5ef43e68433893ec3aaa466 Jan 26 16:58:57 crc kubenswrapper[4856]: I0126 16:58:57.459672 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 13:10:15.341683813 +0000 UTC Jan 26 16:58:57 crc kubenswrapper[4856]: I0126 16:58:57.854500 4856 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 16:58:57 crc kubenswrapper[4856]: I0126 16:58:57.856006 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:58:57 crc kubenswrapper[4856]: I0126 16:58:57.856043 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:58:57 crc kubenswrapper[4856]: I0126 16:58:57.856053 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:58:57 crc kubenswrapper[4856]: I0126 16:58:57.856151 4856 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 26 16:58:57 crc kubenswrapper[4856]: I0126 16:58:57.873056 4856 kubelet_node_status.go:115] "Node was previously registered" node="crc" Jan 26 16:58:57 crc kubenswrapper[4856]: I0126 16:58:57.873332 4856 kubelet_node_status.go:79] "Successfully registered node" node="crc" Jan 26 16:58:57 crc kubenswrapper[4856]: I0126 16:58:57.874553 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:58:57 crc kubenswrapper[4856]: I0126 16:58:57.874584 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:58:57 crc kubenswrapper[4856]: I0126 16:58:57.874593 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:58:57 crc kubenswrapper[4856]: I0126 16:58:57.874608 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:58:57 crc kubenswrapper[4856]: I0126 16:58:57.874626 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:58:57Z","lastTransitionTime":"2026-01-26T16:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:58:57 crc kubenswrapper[4856]: I0126 16:58:57.885618 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-rq622" event={"ID":"7a742e7b-c420-46e3-9e96-e9c744af6124","Type":"ContainerStarted","Data":"ad7222c9b91c0065a545bc1904d9864a5923cc13bfb6617daeb4a965a830f191"} Jan 26 16:58:57 crc kubenswrapper[4856]: I0126 16:58:57.885662 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-rq622" event={"ID":"7a742e7b-c420-46e3-9e96-e9c744af6124","Type":"ContainerStarted","Data":"e0f423c54a0c6586c0aecb6586b15b46e04f0d134b3d8019b80ae941f2d59917"} Jan 26 16:58:57 crc kubenswrapper[4856]: I0126 16:58:57.887173 4856 generic.go:334] "Generic (PLEG): container finished" podID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerID="d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8" exitCode=0 Jan 26 16:58:57 crc kubenswrapper[4856]: I0126 16:58:57.887238 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" event={"ID":"ab5b6f50-172b-4535-a0f9-5d103bcab4e7","Type":"ContainerDied","Data":"d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8"} Jan 26 16:58:57 crc kubenswrapper[4856]: I0126 16:58:57.887282 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" event={"ID":"ab5b6f50-172b-4535-a0f9-5d103bcab4e7","Type":"ContainerStarted","Data":"a1b2fe845f0957cc37219c78a754b5c2b9acc25bf2ef8f7083ca734c4c5c68b9"} Jan 26 16:58:57 crc kubenswrapper[4856]: I0126 16:58:57.888912 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" event={"ID":"63c75ede-5170-4db0-811b-5217ef8d72b3","Type":"ContainerStarted","Data":"da26ddcfc6ccde3c9aabd63bdef3435f9a9eaab8c095bfeef5670c1295576cb0"} Jan 26 16:58:57 crc kubenswrapper[4856]: I0126 16:58:57.888973 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" event={"ID":"63c75ede-5170-4db0-811b-5217ef8d72b3","Type":"ContainerStarted","Data":"54ca3fa13d9e8d442efa93b44a870369f7df3fe7562d77b98528f5c19a751f18"} Jan 26 16:58:57 crc kubenswrapper[4856]: I0126 16:58:57.888984 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" event={"ID":"63c75ede-5170-4db0-811b-5217ef8d72b3","Type":"ContainerStarted","Data":"a4deba88fd2726a90167401cf17b82783a9a1e76e5ef43e68433893ec3aaa466"} Jan 26 16:58:57 crc kubenswrapper[4856]: I0126 16:58:57.894068 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" event={"ID":"ad7b59f9-beb7-49d6-a2d1-e29133e46854","Type":"ContainerStarted","Data":"e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363"} Jan 26 16:58:57 crc kubenswrapper[4856]: I0126 16:58:57.894113 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" event={"ID":"ad7b59f9-beb7-49d6-a2d1-e29133e46854","Type":"ContainerStarted","Data":"90be437457a8690be4cf46ab14042d606cac05b13668a8e2661760045afcc8d8"} Jan 26 16:58:57 crc kubenswrapper[4856]: I0126 16:58:57.907702 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af4930ff9cebd62106545fb3da2dfc93f7b591426fe4c85aa2e637b60c935f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:57Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:57 crc kubenswrapper[4856]: E0126 16:58:57.917517 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17523591-a778-4a97-aeab-8a7a93101850\\\",\\\"systemUUID\\\":\\\"ca45d056-99cb-4442-8a44-7e899628ecb2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:57Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:57 crc kubenswrapper[4856]: I0126 16:58:57.922599 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:58:57 crc kubenswrapper[4856]: I0126 16:58:57.922636 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:58:57 crc kubenswrapper[4856]: I0126 16:58:57.922646 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:58:57 crc kubenswrapper[4856]: I0126 16:58:57.922661 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:58:57 crc kubenswrapper[4856]: I0126 16:58:57.922671 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:58:57Z","lastTransitionTime":"2026-01-26T16:58:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:58:57 crc kubenswrapper[4856]: I0126 16:58:57.968186 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c09748c0963bdb50c9552f249d1135ea53e68c52babd788dad050951cb849cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:57Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:57 crc kubenswrapper[4856]: E0126 16:58:57.982122 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17523591-a778-4a97-aeab-8a7a93101850\\\",\\\"systemUUID\\\":\\\"ca45d056-99cb-4442-8a44-7e899628ecb2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:57Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.091022 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.091080 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.091092 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.091110 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.091123 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:58:58Z","lastTransitionTime":"2026-01-26T16:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.095509 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:58 crc kubenswrapper[4856]: E0126 16:58:58.177586 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17523591-a778-4a97-aeab-8a7a93101850\\\",\\\"systemUUID\\\":\\\"ca45d056-99cb-4442-8a44-7e899628ecb2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.219246 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.219667 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.219751 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.219823 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.219904 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:58:58Z","lastTransitionTime":"2026-01-26T16:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.230906 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.250309 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://627cd6b39fdedd967fffcfc0755439277b2d73016fba1ddc7f5b9d5deba43b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:58 crc kubenswrapper[4856]: E0126 16:58:58.252100 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17523591-a778-4a97-aeab-8a7a93101850\\\",\\\"systemUUID\\\":\\\"ca45d056-99cb-4442-8a44-7e899628ecb2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.262587 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.262917 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.263002 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.263078 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.263136 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:58:58Z","lastTransitionTime":"2026-01-26T16:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:58:58 crc kubenswrapper[4856]: E0126 16:58:58.292370 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17523591-a778-4a97-aeab-8a7a93101850\\\",\\\"systemUUID\\\":\\\"ca45d056-99cb-4442-8a44-7e899628ecb2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:58 crc kubenswrapper[4856]: E0126 16:58:58.292922 4856 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.294586 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.294627 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.294635 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.294647 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.294658 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:58:58Z","lastTransitionTime":"2026-01-26T16:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.295659 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rq622" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a742e7b-c420-46e3-9e96-e9c744af6124\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad7222c9b91c0065a545bc1904d9864a5923cc13bfb6617daeb4a965a830f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8plh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rq622\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.335925 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191ad5d41024c88c1a4fbc30f307dd6340fd55b93b00d22f62209d0e82be286f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2a4d8be409a46ffe5702797e56a80023a72a19fb7dc5c49e5f4984cedc600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.362283 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.376288 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.386704 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.386949 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.387070 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:58:58 crc kubenswrapper[4856]: E0126 16:58:58.387192 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:59:02.387158329 +0000 UTC m=+38.340412340 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.387247 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.387305 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:58:58 crc kubenswrapper[4856]: E0126 16:58:58.387423 4856 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 16:58:58 crc kubenswrapper[4856]: E0126 16:58:58.387471 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 16:59:02.387461918 +0000 UTC m=+38.340715909 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 16:58:58 crc kubenswrapper[4856]: E0126 16:58:58.387581 4856 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 16:58:58 crc kubenswrapper[4856]: E0126 16:58:58.387600 4856 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 16:58:58 crc kubenswrapper[4856]: E0126 16:58:58.387614 4856 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:58:58 crc kubenswrapper[4856]: E0126 16:58:58.387643 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-26 16:59:02.387634652 +0000 UTC m=+38.340888723 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:58:58 crc kubenswrapper[4856]: E0126 16:58:58.388020 4856 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 16:58:58 crc kubenswrapper[4856]: E0126 16:58:58.388108 4856 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 16:58:58 crc kubenswrapper[4856]: E0126 16:58:58.388184 4856 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:58:58 crc kubenswrapper[4856]: E0126 16:58:58.388228 4856 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 16:58:58 crc kubenswrapper[4856]: E0126 16:58:58.388383 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-26 16:59:02.38828727 +0000 UTC m=+38.341541341 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:58:58 crc kubenswrapper[4856]: E0126 16:58:58.388487 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 16:59:02.388464065 +0000 UTC m=+38.341718086 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.393839 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad7b59f9-beb7-49d6-a2d1-e29133e46854\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v2l7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.394092 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:58:58 crc kubenswrapper[4856]: E0126 16:58:58.394171 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.394216 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:58:58 crc kubenswrapper[4856]: E0126 16:58:58.394259 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.394305 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:58:58 crc kubenswrapper[4856]: E0126 16:58:58.394359 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.398405 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.398454 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.398465 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.398483 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.398495 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:58:58Z","lastTransitionTime":"2026-01-26T16:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.423083 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pxh94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.460024 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 03:01:17.974540121 +0000 UTC Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.472479 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63c75ede-5170-4db0-811b-5217ef8d72b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xm9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.486991 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63c75ede-5170-4db0-811b-5217ef8d72b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da26ddcfc6ccde3c9aabd63bdef3435f9a9eaab8c095bfeef5670c1295576cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54ca3fa13d9e8d442efa93b44a870369f7df3fe7562d77b98528f5c19a751f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xm9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.500833 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.500876 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.500896 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.500912 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.500921 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:58:58Z","lastTransitionTime":"2026-01-26T16:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.524161 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af4930ff9cebd62106545fb3da2dfc93f7b591426fe4c85aa2e637b60c935f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.603129 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.603169 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.603182 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.603197 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.603208 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:58:58Z","lastTransitionTime":"2026-01-26T16:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.612324 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c09748c0963bdb50c9552f249d1135ea53e68c52babd788dad050951cb849cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.627992 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.645065 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.657964 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://627cd6b39fdedd967fffcfc0755439277b2d73016fba1ddc7f5b9d5deba43b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.676386 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rq622" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a742e7b-c420-46e3-9e96-e9c744af6124\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad7222c9b91c0065a545bc1904d9864a5923cc13bfb6617daeb4a965a830f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8plh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rq622\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.691939 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191ad5d41024c88c1a4fbc30f307dd6340fd55b93b00d22f62209d0e82be286f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2a4d8be409a46ffe5702797e56a80023a72a19fb7dc5c49e5f4984cedc600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.705488 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.705726 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.705745 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.705764 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.705774 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:58:58Z","lastTransitionTime":"2026-01-26T16:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.708076 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.723497 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.744931 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad7b59f9-beb7-49d6-a2d1-e29133e46854\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v2l7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.771409 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pxh94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.819810 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.819845 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.819857 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.819874 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.819886 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:58:58Z","lastTransitionTime":"2026-01-26T16:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.902761 4856 generic.go:334] "Generic (PLEG): container finished" podID="ad7b59f9-beb7-49d6-a2d1-e29133e46854" containerID="e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363" exitCode=0 Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.902883 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" event={"ID":"ad7b59f9-beb7-49d6-a2d1-e29133e46854","Type":"ContainerDied","Data":"e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363"} Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.916923 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" event={"ID":"ab5b6f50-172b-4535-a0f9-5d103bcab4e7","Type":"ContainerStarted","Data":"4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7"} Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.916987 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" event={"ID":"ab5b6f50-172b-4535-a0f9-5d103bcab4e7","Type":"ContainerStarted","Data":"b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde"} Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.917004 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" event={"ID":"ab5b6f50-172b-4535-a0f9-5d103bcab4e7","Type":"ContainerStarted","Data":"7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3"} Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.917017 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" event={"ID":"ab5b6f50-172b-4535-a0f9-5d103bcab4e7","Type":"ContainerStarted","Data":"e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1"} Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.917029 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" event={"ID":"ab5b6f50-172b-4535-a0f9-5d103bcab4e7","Type":"ContainerStarted","Data":"83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc"} Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.917042 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" event={"ID":"ab5b6f50-172b-4535-a0f9-5d103bcab4e7","Type":"ContainerStarted","Data":"25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b"} Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.918150 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af4930ff9cebd62106545fb3da2dfc93f7b591426fe4c85aa2e637b60c935f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.928401 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.928457 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.928470 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.928494 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.928507 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:58:58Z","lastTransitionTime":"2026-01-26T16:58:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:58:58 crc kubenswrapper[4856]: I0126 16:58:58.932089 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c09748c0963bdb50c9552f249d1135ea53e68c52babd788dad050951cb849cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.031060 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.031110 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.031118 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.031133 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.031142 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:58:59Z","lastTransitionTime":"2026-01-26T16:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.065310 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:59Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.080726 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191ad5d41024c88c1a4fbc30f307dd6340fd55b93b00d22f62209d0e82be286f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2a4d8be409a46ffe5702797e56a80023a72a19fb7dc5c49e5f4984cedc600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:59Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.100906 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:59Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.127799 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:59Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.141930 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.141969 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.141977 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.141994 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.142004 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:58:59Z","lastTransitionTime":"2026-01-26T16:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.151549 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://627cd6b39fdedd967fffcfc0755439277b2d73016fba1ddc7f5b9d5deba43b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:59Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.172181 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rq622" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a742e7b-c420-46e3-9e96-e9c744af6124\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad7222c9b91c0065a545bc1904d9864a5923cc13bfb6617daeb4a965a830f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8plh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rq622\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:59Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.208498 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:59Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.221260 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad7b59f9-beb7-49d6-a2d1-e29133e46854\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v2l7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:59Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.240864 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pxh94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:59Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.244834 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.244858 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.244865 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.244879 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.244889 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:58:59Z","lastTransitionTime":"2026-01-26T16:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.253913 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63c75ede-5170-4db0-811b-5217ef8d72b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da26ddcfc6ccde3c9aabd63bdef3435f9a9eaab8c095bfeef5670c1295576cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54ca3fa13d9e8d442efa93b44a870369f7df3fe7562d77b98528f5c19a751f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xm9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:59Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.347931 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.348374 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.348387 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.348406 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.348419 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:58:59Z","lastTransitionTime":"2026-01-26T16:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.451431 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.451461 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.451472 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.451486 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.451496 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:58:59Z","lastTransitionTime":"2026-01-26T16:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.460145 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-30 13:25:07.820004316 +0000 UTC Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.553263 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.553285 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.553292 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.553306 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.553315 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:58:59Z","lastTransitionTime":"2026-01-26T16:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.655679 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.655705 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.655713 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.655727 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.655736 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:58:59Z","lastTransitionTime":"2026-01-26T16:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.757845 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.757889 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.757900 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.757917 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.757929 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:58:59Z","lastTransitionTime":"2026-01-26T16:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.859277 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.859302 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.859310 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.859323 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.859331 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:58:59Z","lastTransitionTime":"2026-01-26T16:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.947561 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" event={"ID":"ad7b59f9-beb7-49d6-a2d1-e29133e46854","Type":"ContainerStarted","Data":"fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196"} Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.960005 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63c75ede-5170-4db0-811b-5217ef8d72b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da26ddcfc6ccde3c9aabd63bdef3435f9a9eaab8c095bfeef5670c1295576cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54ca3fa13d9e8d442efa93b44a870369f7df3fe7562d77b98528f5c19a751f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xm9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:59Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.961469 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.961515 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.961545 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.961563 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.961574 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:58:59Z","lastTransitionTime":"2026-01-26T16:58:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.973439 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af4930ff9cebd62106545fb3da2dfc93f7b591426fe4c85aa2e637b60c935f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:59Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.984173 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c09748c0963bdb50c9552f249d1135ea53e68c52babd788dad050951cb849cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:59Z is after 2025-08-24T17:21:41Z" Jan 26 16:58:59 crc kubenswrapper[4856]: I0126 16:58:59.997036 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:58:59Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.010227 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191ad5d41024c88c1a4fbc30f307dd6340fd55b93b00d22f62209d0e82be286f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2a4d8be409a46ffe5702797e56a80023a72a19fb7dc5c49e5f4984cedc600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:00Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.028695 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:00Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.041789 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:00Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.052414 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://627cd6b39fdedd967fffcfc0755439277b2d73016fba1ddc7f5b9d5deba43b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:00Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.097867 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rq622" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a742e7b-c420-46e3-9e96-e9c744af6124\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad7222c9b91c0065a545bc1904d9864a5923cc13bfb6617daeb4a965a830f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8plh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rq622\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:00Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.114748 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.114791 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.114801 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.114816 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.114826 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:00Z","lastTransitionTime":"2026-01-26T16:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.122279 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad7b59f9-beb7-49d6-a2d1-e29133e46854\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v2l7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:00Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.140780 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pxh94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:00Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.156672 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:00Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.217658 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.217684 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.217692 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.217705 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.217714 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:00Z","lastTransitionTime":"2026-01-26T16:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.320225 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.320263 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.320274 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.320290 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.320305 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:00Z","lastTransitionTime":"2026-01-26T16:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.394749 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.394811 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.394767 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:59:00 crc kubenswrapper[4856]: E0126 16:59:00.394894 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:59:00 crc kubenswrapper[4856]: E0126 16:59:00.395004 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:59:00 crc kubenswrapper[4856]: E0126 16:59:00.395078 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.423254 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.423292 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.423301 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.423318 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.423334 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:00Z","lastTransitionTime":"2026-01-26T16:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.460841 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 12:53:02.635599378 +0000 UTC Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.525868 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.525904 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.525912 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.525926 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.525938 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:00Z","lastTransitionTime":"2026-01-26T16:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.628710 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.629042 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.629056 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.629075 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.629087 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:00Z","lastTransitionTime":"2026-01-26T16:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.731918 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.731988 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.732001 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.732019 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.732031 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:00Z","lastTransitionTime":"2026-01-26T16:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.834459 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.834518 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.834556 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.834583 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.834601 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:00Z","lastTransitionTime":"2026-01-26T16:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.937271 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.937312 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.937324 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.937346 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.937359 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:00Z","lastTransitionTime":"2026-01-26T16:59:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.951553 4856 generic.go:334] "Generic (PLEG): container finished" podID="ad7b59f9-beb7-49d6-a2d1-e29133e46854" containerID="fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196" exitCode=0 Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.951610 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" event={"ID":"ad7b59f9-beb7-49d6-a2d1-e29133e46854","Type":"ContainerDied","Data":"fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196"} Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.964575 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63c75ede-5170-4db0-811b-5217ef8d72b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da26ddcfc6ccde3c9aabd63bdef3435f9a9eaab8c095bfeef5670c1295576cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54ca3fa13d9e8d442efa93b44a870369f7df3fe7562d77b98528f5c19a751f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xm9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:00Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.984675 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af4930ff9cebd62106545fb3da2dfc93f7b591426fe4c85aa2e637b60c935f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:00Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:00 crc kubenswrapper[4856]: I0126 16:59:00.996934 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c09748c0963bdb50c9552f249d1135ea53e68c52babd788dad050951cb849cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:00Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.011635 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:01Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.023505 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://627cd6b39fdedd967fffcfc0755439277b2d73016fba1ddc7f5b9d5deba43b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:01Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.036820 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rq622" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a742e7b-c420-46e3-9e96-e9c744af6124\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad7222c9b91c0065a545bc1904d9864a5923cc13bfb6617daeb4a965a830f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8plh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rq622\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:01Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.041848 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.041878 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.041886 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.041901 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.041911 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:01Z","lastTransitionTime":"2026-01-26T16:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.051369 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191ad5d41024c88c1a4fbc30f307dd6340fd55b93b00d22f62209d0e82be286f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2a4d8be409a46ffe5702797e56a80023a72a19fb7dc5c49e5f4984cedc600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:01Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.064663 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:01Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.080151 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:01Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.097142 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:01Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.111801 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad7b59f9-beb7-49d6-a2d1-e29133e46854\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v2l7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:01Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.131191 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pxh94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:01Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.144623 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.144660 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.144672 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.144689 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.144701 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:01Z","lastTransitionTime":"2026-01-26T16:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.246664 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.246702 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.246711 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.246728 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.246740 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:01Z","lastTransitionTime":"2026-01-26T16:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.349610 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.349653 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.349665 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.349695 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.349723 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:01Z","lastTransitionTime":"2026-01-26T16:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.451780 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.451831 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.451852 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.451873 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.451896 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:01Z","lastTransitionTime":"2026-01-26T16:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.461158 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 06:39:35.931342519 +0000 UTC Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.554456 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.554703 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.554712 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.554726 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.554734 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:01Z","lastTransitionTime":"2026-01-26T16:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.658989 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.659015 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.659026 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.659042 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.659052 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:01Z","lastTransitionTime":"2026-01-26T16:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.762580 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.762638 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.762653 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.762676 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.762692 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:01Z","lastTransitionTime":"2026-01-26T16:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.865089 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.865133 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.865144 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.865162 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.865172 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:01Z","lastTransitionTime":"2026-01-26T16:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.958396 4856 generic.go:334] "Generic (PLEG): container finished" podID="ad7b59f9-beb7-49d6-a2d1-e29133e46854" containerID="79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993" exitCode=0 Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.958455 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" event={"ID":"ad7b59f9-beb7-49d6-a2d1-e29133e46854","Type":"ContainerDied","Data":"79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993"} Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.969841 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.969880 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.969891 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.969909 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.969918 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:01Z","lastTransitionTime":"2026-01-26T16:59:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.971038 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63c75ede-5170-4db0-811b-5217ef8d72b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da26ddcfc6ccde3c9aabd63bdef3435f9a9eaab8c095bfeef5670c1295576cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54ca3fa13d9e8d442efa93b44a870369f7df3fe7562d77b98528f5c19a751f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xm9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:01Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.971975 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" event={"ID":"ab5b6f50-172b-4535-a0f9-5d103bcab4e7","Type":"ContainerStarted","Data":"11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a"} Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.983451 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"c823748751e9938f10fab08e33b5fffff5a6d15961ce028204131b7b69a56c18"} Jan 26 16:59:01 crc kubenswrapper[4856]: I0126 16:59:01.987055 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c09748c0963bdb50c9552f249d1135ea53e68c52babd788dad050951cb849cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:01Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.006971 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:02Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.022167 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af4930ff9cebd62106545fb3da2dfc93f7b591426fe4c85aa2e637b60c935f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:02Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.035936 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191ad5d41024c88c1a4fbc30f307dd6340fd55b93b00d22f62209d0e82be286f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2a4d8be409a46ffe5702797e56a80023a72a19fb7dc5c49e5f4984cedc600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:02Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.049148 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:02Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.062314 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:02Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.072639 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.072668 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.072679 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.072693 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.072703 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:02Z","lastTransitionTime":"2026-01-26T16:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.073881 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://627cd6b39fdedd967fffcfc0755439277b2d73016fba1ddc7f5b9d5deba43b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:02Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.090197 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rq622" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a742e7b-c420-46e3-9e96-e9c744af6124\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad7222c9b91c0065a545bc1904d9864a5923cc13bfb6617daeb4a965a830f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8plh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rq622\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:02Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.110726 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pxh94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:02Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.124102 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:02Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.142777 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad7b59f9-beb7-49d6-a2d1-e29133e46854\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v2l7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:02Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.159444 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad7b59f9-beb7-49d6-a2d1-e29133e46854\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v2l7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:02Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.178792 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pxh94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:02Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.185337 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.185373 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.185384 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.185399 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.185409 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:02Z","lastTransitionTime":"2026-01-26T16:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.193465 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:02Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.208065 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63c75ede-5170-4db0-811b-5217ef8d72b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da26ddcfc6ccde3c9aabd63bdef3435f9a9eaab8c095bfeef5670c1295576cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54ca3fa13d9e8d442efa93b44a870369f7df3fe7562d77b98528f5c19a751f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xm9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:02Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.220353 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af4930ff9cebd62106545fb3da2dfc93f7b591426fe4c85aa2e637b60c935f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:02Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.234964 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c09748c0963bdb50c9552f249d1135ea53e68c52babd788dad050951cb849cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:02Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.250263 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:02Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.270693 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191ad5d41024c88c1a4fbc30f307dd6340fd55b93b00d22f62209d0e82be286f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2a4d8be409a46ffe5702797e56a80023a72a19fb7dc5c49e5f4984cedc600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:02Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.284494 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c823748751e9938f10fab08e33b5fffff5a6d15961ce028204131b7b69a56c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:02Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.288036 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.288072 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.288081 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.288096 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.288106 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:02Z","lastTransitionTime":"2026-01-26T16:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.297279 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:02Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.307478 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://627cd6b39fdedd967fffcfc0755439277b2d73016fba1ddc7f5b9d5deba43b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:02Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.319603 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rq622" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a742e7b-c420-46e3-9e96-e9c744af6124\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad7222c9b91c0065a545bc1904d9864a5923cc13bfb6617daeb4a965a830f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8plh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rq622\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:02Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.390171 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.390225 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.390236 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.390255 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.390274 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:02Z","lastTransitionTime":"2026-01-26T16:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.394373 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.394399 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.394402 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:59:02 crc kubenswrapper[4856]: E0126 16:59:02.394474 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:59:02 crc kubenswrapper[4856]: E0126 16:59:02.394567 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:59:02 crc kubenswrapper[4856]: E0126 16:59:02.394618 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.461591 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 04:40:00.515652777 +0000 UTC Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.479269 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.479503 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:59:02 crc kubenswrapper[4856]: E0126 16:59:02.479554 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:59:10.479496688 +0000 UTC m=+46.432750669 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:59:02 crc kubenswrapper[4856]: E0126 16:59:02.479668 4856 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 16:59:02 crc kubenswrapper[4856]: E0126 16:59:02.479876 4856 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 16:59:02 crc kubenswrapper[4856]: E0126 16:59:02.479904 4856 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 16:59:02 crc kubenswrapper[4856]: E0126 16:59:02.479924 4856 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:59:02 crc kubenswrapper[4856]: E0126 16:59:02.479883 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 16:59:10.479869739 +0000 UTC m=+46.433123950 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 16:59:02 crc kubenswrapper[4856]: E0126 16:59:02.480015 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-26 16:59:10.479992532 +0000 UTC m=+46.433246513 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.479746 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.480071 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.480102 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:59:02 crc kubenswrapper[4856]: E0126 16:59:02.480244 4856 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 16:59:02 crc kubenswrapper[4856]: E0126 16:59:02.480280 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 16:59:10.48027272 +0000 UTC m=+46.433526701 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 16:59:02 crc kubenswrapper[4856]: E0126 16:59:02.480405 4856 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 16:59:02 crc kubenswrapper[4856]: E0126 16:59:02.480467 4856 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 16:59:02 crc kubenswrapper[4856]: E0126 16:59:02.480485 4856 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:59:02 crc kubenswrapper[4856]: E0126 16:59:02.480605 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-26 16:59:10.480570738 +0000 UTC m=+46.433824719 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.492557 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.492622 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.492647 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.492677 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.492702 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:02Z","lastTransitionTime":"2026-01-26T16:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.595657 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.595711 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.595723 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.595742 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.595756 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:02Z","lastTransitionTime":"2026-01-26T16:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.698708 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.698744 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.698754 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.698768 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.698777 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:02Z","lastTransitionTime":"2026-01-26T16:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.801263 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.801315 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.801325 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.801341 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.801350 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:02Z","lastTransitionTime":"2026-01-26T16:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.904491 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.904551 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.904564 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.904583 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.904595 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:02Z","lastTransitionTime":"2026-01-26T16:59:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.989089 4856 generic.go:334] "Generic (PLEG): container finished" podID="ad7b59f9-beb7-49d6-a2d1-e29133e46854" containerID="c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1" exitCode=0 Jan 26 16:59:02 crc kubenswrapper[4856]: I0126 16:59:02.989159 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" event={"ID":"ad7b59f9-beb7-49d6-a2d1-e29133e46854","Type":"ContainerDied","Data":"c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1"} Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.005069 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:03Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.007267 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.007309 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.007320 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.007338 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.007350 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:03Z","lastTransitionTime":"2026-01-26T16:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.022839 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad7b59f9-beb7-49d6-a2d1-e29133e46854\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v2l7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:03Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.043467 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pxh94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:03Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.056794 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63c75ede-5170-4db0-811b-5217ef8d72b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da26ddcfc6ccde3c9aabd63bdef3435f9a9eaab8c095bfeef5670c1295576cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54ca3fa13d9e8d442efa93b44a870369f7df3fe7562d77b98528f5c19a751f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xm9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:03Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.070876 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af4930ff9cebd62106545fb3da2dfc93f7b591426fe4c85aa2e637b60c935f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:03Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.084351 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c09748c0963bdb50c9552f249d1135ea53e68c52babd788dad050951cb849cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:03Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.099626 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:03Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.110754 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.110799 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.110856 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.110875 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.110887 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:03Z","lastTransitionTime":"2026-01-26T16:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.114249 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191ad5d41024c88c1a4fbc30f307dd6340fd55b93b00d22f62209d0e82be286f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2a4d8be409a46ffe5702797e56a80023a72a19fb7dc5c49e5f4984cedc600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:03Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.129196 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c823748751e9938f10fab08e33b5fffff5a6d15961ce028204131b7b69a56c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:03Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.142121 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:03Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.154826 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://627cd6b39fdedd967fffcfc0755439277b2d73016fba1ddc7f5b9d5deba43b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:03Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.170185 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rq622" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a742e7b-c420-46e3-9e96-e9c744af6124\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad7222c9b91c0065a545bc1904d9864a5923cc13bfb6617daeb4a965a830f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8plh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rq622\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:03Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.214074 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.214111 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.214121 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.214136 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.214146 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:03Z","lastTransitionTime":"2026-01-26T16:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.316848 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.316891 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.316900 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.316917 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.316928 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:03Z","lastTransitionTime":"2026-01-26T16:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.420881 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.420932 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.420945 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.420962 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.420982 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:03Z","lastTransitionTime":"2026-01-26T16:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.461956 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 03:03:49.251194883 +0000 UTC Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.523855 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.523895 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.523905 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.523921 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.523931 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:03Z","lastTransitionTime":"2026-01-26T16:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.626874 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.626914 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.626928 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.626946 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.626959 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:03Z","lastTransitionTime":"2026-01-26T16:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.730618 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.730683 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.730699 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.730724 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.730743 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:03Z","lastTransitionTime":"2026-01-26T16:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.839284 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.839344 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.839364 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.839392 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.839410 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:03Z","lastTransitionTime":"2026-01-26T16:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.941654 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.941729 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.941755 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.941790 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.941814 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:03Z","lastTransitionTime":"2026-01-26T16:59:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.995417 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" event={"ID":"ad7b59f9-beb7-49d6-a2d1-e29133e46854","Type":"ContainerStarted","Data":"62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea"} Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.997866 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" event={"ID":"ab5b6f50-172b-4535-a0f9-5d103bcab4e7","Type":"ContainerStarted","Data":"36e1e2c0e79e963e1ad9b28e9b6e7d69c6b6df040359eef2630b4aeb32109f38"} Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.998621 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.998716 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:59:03 crc kubenswrapper[4856]: I0126 16:59:03.998734 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.016499 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191ad5d41024c88c1a4fbc30f307dd6340fd55b93b00d22f62209d0e82be286f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2a4d8be409a46ffe5702797e56a80023a72a19fb7dc5c49e5f4984cedc600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:04Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.030850 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c823748751e9938f10fab08e33b5fffff5a6d15961ce028204131b7b69a56c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:04Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.046351 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:04Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.062567 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://627cd6b39fdedd967fffcfc0755439277b2d73016fba1ddc7f5b9d5deba43b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:04Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.078703 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.078904 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.078920 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.078928 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.078941 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.078950 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:04Z","lastTransitionTime":"2026-01-26T16:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.079081 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rq622" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a742e7b-c420-46e3-9e96-e9c744af6124\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad7222c9b91c0065a545bc1904d9864a5923cc13bfb6617daeb4a965a830f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8plh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rq622\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:04Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.091878 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:04Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.118291 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad7b59f9-beb7-49d6-a2d1-e29133e46854\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v2l7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:04Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.132132 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.135611 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pxh94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:04Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.147682 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63c75ede-5170-4db0-811b-5217ef8d72b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da26ddcfc6ccde3c9aabd63bdef3435f9a9eaab8c095bfeef5670c1295576cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54ca3fa13d9e8d442efa93b44a870369f7df3fe7562d77b98528f5c19a751f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xm9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:04Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.157823 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af4930ff9cebd62106545fb3da2dfc93f7b591426fe4c85aa2e637b60c935f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:04Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.174776 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c09748c0963bdb50c9552f249d1135ea53e68c52babd788dad050951cb849cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:04Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.180972 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.181016 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.181027 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.181046 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.181058 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:04Z","lastTransitionTime":"2026-01-26T16:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.194138 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:04Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.206186 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af4930ff9cebd62106545fb3da2dfc93f7b591426fe4c85aa2e637b60c935f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:04Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.225676 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c09748c0963bdb50c9552f249d1135ea53e68c52babd788dad050951cb849cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:04Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.242935 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:04Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.265374 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191ad5d41024c88c1a4fbc30f307dd6340fd55b93b00d22f62209d0e82be286f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2a4d8be409a46ffe5702797e56a80023a72a19fb7dc5c49e5f4984cedc600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:04Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.281471 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c823748751e9938f10fab08e33b5fffff5a6d15961ce028204131b7b69a56c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:04Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.283655 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.283708 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.283721 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.283742 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.283755 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:04Z","lastTransitionTime":"2026-01-26T16:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.295955 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:04Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.309272 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://627cd6b39fdedd967fffcfc0755439277b2d73016fba1ddc7f5b9d5deba43b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:04Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.327583 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rq622" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a742e7b-c420-46e3-9e96-e9c744af6124\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad7222c9b91c0065a545bc1904d9864a5923cc13bfb6617daeb4a965a830f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8plh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rq622\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:04Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.340955 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:04Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.355477 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad7b59f9-beb7-49d6-a2d1-e29133e46854\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v2l7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:04Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.375865 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36e1e2c0e79e963e1ad9b28e9b6e7d69c6b6df040359eef2630b4aeb32109f38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pxh94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:04Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.386125 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.386172 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.386183 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.386198 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.386208 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:04Z","lastTransitionTime":"2026-01-26T16:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.387779 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63c75ede-5170-4db0-811b-5217ef8d72b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da26ddcfc6ccde3c9aabd63bdef3435f9a9eaab8c095bfeef5670c1295576cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54ca3fa13d9e8d442efa93b44a870369f7df3fe7562d77b98528f5c19a751f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xm9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:04Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.395025 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.395105 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:59:04 crc kubenswrapper[4856]: E0126 16:59:04.395159 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.395286 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:59:04 crc kubenswrapper[4856]: E0126 16:59:04.395438 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:59:04 crc kubenswrapper[4856]: E0126 16:59:04.395596 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.462608 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 15:18:50.096384286 +0000 UTC Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.488820 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.488864 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.488874 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.488891 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.488901 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:04Z","lastTransitionTime":"2026-01-26T16:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.592541 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.592589 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.592604 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.592623 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.592638 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:04Z","lastTransitionTime":"2026-01-26T16:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.700053 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.700090 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.700099 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.700114 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.700124 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:04Z","lastTransitionTime":"2026-01-26T16:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.802110 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.802138 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.802145 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.802157 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.802166 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:04Z","lastTransitionTime":"2026-01-26T16:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.904101 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.904157 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.904167 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.904184 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:04 crc kubenswrapper[4856]: I0126 16:59:04.904194 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:04Z","lastTransitionTime":"2026-01-26T16:59:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.028159 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.028192 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.028201 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.028215 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.028225 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:05Z","lastTransitionTime":"2026-01-26T16:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.132276 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.132317 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.132326 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.132343 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.132354 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:05Z","lastTransitionTime":"2026-01-26T16:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.235501 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.235571 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.235582 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.235599 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.235611 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:05Z","lastTransitionTime":"2026-01-26T16:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.337938 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.338264 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.338277 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.338290 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.338299 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:05Z","lastTransitionTime":"2026-01-26T16:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.440785 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.440822 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.440834 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.440856 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.440868 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:05Z","lastTransitionTime":"2026-01-26T16:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.462972 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 09:50:47.731554375 +0000 UTC Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.464939 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63c75ede-5170-4db0-811b-5217ef8d72b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da26ddcfc6ccde3c9aabd63bdef3435f9a9eaab8c095bfeef5670c1295576cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54ca3fa13d9e8d442efa93b44a870369f7df3fe7562d77b98528f5c19a751f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xm9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:05Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.478660 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af4930ff9cebd62106545fb3da2dfc93f7b591426fe4c85aa2e637b60c935f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:05Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.493667 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c09748c0963bdb50c9552f249d1135ea53e68c52babd788dad050951cb849cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:05Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.508786 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:05Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.522418 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rq622" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a742e7b-c420-46e3-9e96-e9c744af6124\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad7222c9b91c0065a545bc1904d9864a5923cc13bfb6617daeb4a965a830f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8plh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rq622\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:05Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.537485 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191ad5d41024c88c1a4fbc30f307dd6340fd55b93b00d22f62209d0e82be286f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2a4d8be409a46ffe5702797e56a80023a72a19fb7dc5c49e5f4984cedc600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:05Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.543208 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.543253 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.543267 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.543287 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.543300 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:05Z","lastTransitionTime":"2026-01-26T16:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.550392 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c823748751e9938f10fab08e33b5fffff5a6d15961ce028204131b7b69a56c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:05Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.567196 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:05Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.578581 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://627cd6b39fdedd967fffcfc0755439277b2d73016fba1ddc7f5b9d5deba43b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:05Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.595235 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:05Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.613276 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad7b59f9-beb7-49d6-a2d1-e29133e46854\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v2l7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:05Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.633400 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36e1e2c0e79e963e1ad9b28e9b6e7d69c6b6df040359eef2630b4aeb32109f38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pxh94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:05Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.645749 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.646004 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.646068 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.646131 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.646209 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:05Z","lastTransitionTime":"2026-01-26T16:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.748876 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.748912 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.748923 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.748937 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.748946 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:05Z","lastTransitionTime":"2026-01-26T16:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.851541 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.851600 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.851610 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.851662 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.851672 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:05Z","lastTransitionTime":"2026-01-26T16:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.954297 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.954379 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.954411 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.954441 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:05 crc kubenswrapper[4856]: I0126 16:59:05.954462 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:05Z","lastTransitionTime":"2026-01-26T16:59:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.056667 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.056706 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.056714 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.056732 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.056742 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:06Z","lastTransitionTime":"2026-01-26T16:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.158971 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.159012 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.159044 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.159062 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.159071 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:06Z","lastTransitionTime":"2026-01-26T16:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.262007 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.262085 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.262106 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.262129 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.262142 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:06Z","lastTransitionTime":"2026-01-26T16:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.365145 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.365216 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.365239 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.365272 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.365297 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:06Z","lastTransitionTime":"2026-01-26T16:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.394503 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.394955 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.394974 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:59:06 crc kubenswrapper[4856]: E0126 16:59:06.395106 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:59:06 crc kubenswrapper[4856]: E0126 16:59:06.395279 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:59:06 crc kubenswrapper[4856]: E0126 16:59:06.395551 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.463648 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 17:17:55.081671222 +0000 UTC Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.467992 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.468071 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.468118 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.468142 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.468157 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:06Z","lastTransitionTime":"2026-01-26T16:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.570860 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.570922 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.570935 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.570953 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.570965 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:06Z","lastTransitionTime":"2026-01-26T16:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.673988 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.674042 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.674054 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.674071 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.674081 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:06Z","lastTransitionTime":"2026-01-26T16:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.776709 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.776761 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.776772 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.776795 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.776808 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:06Z","lastTransitionTime":"2026-01-26T16:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.879264 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.879323 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.879344 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.879366 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.879380 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:06Z","lastTransitionTime":"2026-01-26T16:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.982588 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.982623 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.982634 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.982651 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:06 crc kubenswrapper[4856]: I0126 16:59:06.982663 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:06Z","lastTransitionTime":"2026-01-26T16:59:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.008266 4856 generic.go:334] "Generic (PLEG): container finished" podID="ad7b59f9-beb7-49d6-a2d1-e29133e46854" containerID="62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea" exitCode=0 Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.008311 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" event={"ID":"ad7b59f9-beb7-49d6-a2d1-e29133e46854","Type":"ContainerDied","Data":"62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea"} Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.026006 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:07Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.042735 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad7b59f9-beb7-49d6-a2d1-e29133e46854\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v2l7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:07Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.066661 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36e1e2c0e79e963e1ad9b28e9b6e7d69c6b6df040359eef2630b4aeb32109f38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pxh94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:07Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.080421 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63c75ede-5170-4db0-811b-5217ef8d72b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da26ddcfc6ccde3c9aabd63bdef3435f9a9eaab8c095bfeef5670c1295576cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54ca3fa13d9e8d442efa93b44a870369f7df3fe7562d77b98528f5c19a751f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xm9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:07Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.084888 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.084919 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.084928 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.084941 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.084954 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:07Z","lastTransitionTime":"2026-01-26T16:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.090868 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af4930ff9cebd62106545fb3da2dfc93f7b591426fe4c85aa2e637b60c935f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:07Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.106200 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c09748c0963bdb50c9552f249d1135ea53e68c52babd788dad050951cb849cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:07Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.119340 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:07Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.131344 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://627cd6b39fdedd967fffcfc0755439277b2d73016fba1ddc7f5b9d5deba43b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:07Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.144351 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rq622" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a742e7b-c420-46e3-9e96-e9c744af6124\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad7222c9b91c0065a545bc1904d9864a5923cc13bfb6617daeb4a965a830f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8plh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rq622\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:07Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.158581 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191ad5d41024c88c1a4fbc30f307dd6340fd55b93b00d22f62209d0e82be286f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2a4d8be409a46ffe5702797e56a80023a72a19fb7dc5c49e5f4984cedc600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:07Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.171601 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c823748751e9938f10fab08e33b5fffff5a6d15961ce028204131b7b69a56c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:07Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.186590 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.186619 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.186628 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.186642 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.186651 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:07Z","lastTransitionTime":"2026-01-26T16:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.202292 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:07Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.289159 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.289197 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.289206 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.289223 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.289234 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:07Z","lastTransitionTime":"2026-01-26T16:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.393128 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.393176 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.393188 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.393206 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.393216 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:07Z","lastTransitionTime":"2026-01-26T16:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.464889 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-16 13:41:09.996253405 +0000 UTC Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.497395 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.497468 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.497486 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.497512 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.497571 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:07Z","lastTransitionTime":"2026-01-26T16:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.599920 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.599975 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.599987 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.600005 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.600015 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:07Z","lastTransitionTime":"2026-01-26T16:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.702635 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.702680 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.702688 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.702702 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.702715 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:07Z","lastTransitionTime":"2026-01-26T16:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.805093 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.805389 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.805482 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.805616 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.805821 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:07Z","lastTransitionTime":"2026-01-26T16:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.908671 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.909105 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.909194 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.909284 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:07 crc kubenswrapper[4856]: I0126 16:59:07.909375 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:07Z","lastTransitionTime":"2026-01-26T16:59:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.016897 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.016927 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.016936 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.016952 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.016963 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:08Z","lastTransitionTime":"2026-01-26T16:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.075075 4856 generic.go:334] "Generic (PLEG): container finished" podID="ad7b59f9-beb7-49d6-a2d1-e29133e46854" containerID="249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8" exitCode=0 Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.075308 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" event={"ID":"ad7b59f9-beb7-49d6-a2d1-e29133e46854","Type":"ContainerDied","Data":"249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8"} Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.100277 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36e1e2c0e79e963e1ad9b28e9b6e7d69c6b6df040359eef2630b4aeb32109f38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pxh94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:08Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.114943 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:08Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.127753 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.127812 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.127834 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.127858 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.127874 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:08Z","lastTransitionTime":"2026-01-26T16:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.133757 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad7b59f9-beb7-49d6-a2d1-e29133e46854\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v2l7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:08Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.144768 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63c75ede-5170-4db0-811b-5217ef8d72b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da26ddcfc6ccde3c9aabd63bdef3435f9a9eaab8c095bfeef5670c1295576cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54ca3fa13d9e8d442efa93b44a870369f7df3fe7562d77b98528f5c19a751f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xm9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:08Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.160160 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c09748c0963bdb50c9552f249d1135ea53e68c52babd788dad050951cb849cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:08Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.175853 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:08Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.188722 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af4930ff9cebd62106545fb3da2dfc93f7b591426fe4c85aa2e637b60c935f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:08Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.203194 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191ad5d41024c88c1a4fbc30f307dd6340fd55b93b00d22f62209d0e82be286f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2a4d8be409a46ffe5702797e56a80023a72a19fb7dc5c49e5f4984cedc600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:08Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.216091 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c823748751e9938f10fab08e33b5fffff5a6d15961ce028204131b7b69a56c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:08Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.228396 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:08Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.231215 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.231257 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.231266 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.231283 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.231292 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:08Z","lastTransitionTime":"2026-01-26T16:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.241516 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://627cd6b39fdedd967fffcfc0755439277b2d73016fba1ddc7f5b9d5deba43b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:08Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.259614 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rq622" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a742e7b-c420-46e3-9e96-e9c744af6124\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad7222c9b91c0065a545bc1904d9864a5923cc13bfb6617daeb4a965a830f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8plh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rq622\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:08Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.333258 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.333303 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.333324 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.333341 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.333364 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:08Z","lastTransitionTime":"2026-01-26T16:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.342965 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.343005 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.343015 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.343030 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.343040 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:08Z","lastTransitionTime":"2026-01-26T16:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:08 crc kubenswrapper[4856]: E0126 16:59:08.355151 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17523591-a778-4a97-aeab-8a7a93101850\\\",\\\"systemUUID\\\":\\\"ca45d056-99cb-4442-8a44-7e899628ecb2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:08Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.359266 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.359329 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.359351 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.359374 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.359396 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:08Z","lastTransitionTime":"2026-01-26T16:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:08 crc kubenswrapper[4856]: E0126 16:59:08.372254 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17523591-a778-4a97-aeab-8a7a93101850\\\",\\\"systemUUID\\\":\\\"ca45d056-99cb-4442-8a44-7e899628ecb2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:08Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.376502 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.376570 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.376588 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.376611 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.376628 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:08Z","lastTransitionTime":"2026-01-26T16:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:08 crc kubenswrapper[4856]: E0126 16:59:08.390938 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17523591-a778-4a97-aeab-8a7a93101850\\\",\\\"systemUUID\\\":\\\"ca45d056-99cb-4442-8a44-7e899628ecb2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:08Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.394539 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.394588 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.394557 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:59:08 crc kubenswrapper[4856]: E0126 16:59:08.394869 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:59:08 crc kubenswrapper[4856]: E0126 16:59:08.395006 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:59:08 crc kubenswrapper[4856]: E0126 16:59:08.395123 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.397632 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.397680 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.397692 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.397709 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.397721 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:08Z","lastTransitionTime":"2026-01-26T16:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:08 crc kubenswrapper[4856]: E0126 16:59:08.430325 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17523591-a778-4a97-aeab-8a7a93101850\\\",\\\"systemUUID\\\":\\\"ca45d056-99cb-4442-8a44-7e899628ecb2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:08Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.431099 4856 scope.go:117] "RemoveContainer" containerID="3f07438e20bdf71c752bb661084c835341999c561bb5442d75e177223881276f" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.432478 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.440133 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.440165 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.440173 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.440189 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.440199 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:08Z","lastTransitionTime":"2026-01-26T16:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:08 crc kubenswrapper[4856]: E0126 16:59:08.456654 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17523591-a778-4a97-aeab-8a7a93101850\\\",\\\"systemUUID\\\":\\\"ca45d056-99cb-4442-8a44-7e899628ecb2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:08Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:08 crc kubenswrapper[4856]: E0126 16:59:08.456814 4856 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.459663 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.459703 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.459713 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.459740 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.459755 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:08Z","lastTransitionTime":"2026-01-26T16:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.465362 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 03:50:12.578301114 +0000 UTC Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.571866 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.571930 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.571939 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.571954 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.571966 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:08Z","lastTransitionTime":"2026-01-26T16:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.674408 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.674455 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.674469 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.674486 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.674497 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:08Z","lastTransitionTime":"2026-01-26T16:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.776939 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.776992 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.777005 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.777027 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.777041 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:08Z","lastTransitionTime":"2026-01-26T16:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.880085 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.880135 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.880148 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.880193 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.880207 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:08Z","lastTransitionTime":"2026-01-26T16:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.986132 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.986386 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.986398 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.986416 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:08 crc kubenswrapper[4856]: I0126 16:59:08.986428 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:08Z","lastTransitionTime":"2026-01-26T16:59:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.097907 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.097958 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.097970 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.097988 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.098004 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:09Z","lastTransitionTime":"2026-01-26T16:59:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.102514 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" event={"ID":"ad7b59f9-beb7-49d6-a2d1-e29133e46854","Type":"ContainerStarted","Data":"fc53a6a6d4222ee5e1ac29fd5957d8cdf3fd42de72c68c85329374ce7afc4004"} Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.105451 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.108188 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b03a01b651d9f66da4dd1f6e0d29ad97c0d6ae46b644c3d997d8dc99476706df"} Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.108477 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.118290 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:09Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.133837 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad7b59f9-beb7-49d6-a2d1-e29133e46854\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53a6a6d4222ee5e1ac29fd5957d8cdf3fd42de72c68c85329374ce7afc4004\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v2l7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:09Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.152853 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36e1e2c0e79e963e1ad9b28e9b6e7d69c6b6df040359eef2630b4aeb32109f38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pxh94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:09Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.165989 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63c75ede-5170-4db0-811b-5217ef8d72b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da26ddcfc6ccde3c9aabd63bdef3435f9a9eaab8c095bfeef5670c1295576cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54ca3fa13d9e8d442efa93b44a870369f7df3fe7562d77b98528f5c19a751f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xm9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:09Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.177584 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af4930ff9cebd62106545fb3da2dfc93f7b591426fe4c85aa2e637b60c935f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:09Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.191839 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c09748c0963bdb50c9552f249d1135ea53e68c52babd788dad050951cb849cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:09Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.198202 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v7579"] Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.198830 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v7579" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.199947 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.199983 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.199993 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.200008 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.200018 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:09Z","lastTransitionTime":"2026-01-26T16:59:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.200680 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.201311 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.205461 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:09Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.224818 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ecd87a-c5db-446d-ad3e-cfabbd648c1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89975b8f9428f81ab5d3fb48ced5dd9c837bea2feea3b89f5f7ff8d7d5d15b3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d432159a07ba20bf95f058bf8a597d67cbd8d852519bed035694f8ba3d8ec4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59ba0f131eee14df389391e46bc772894e49e976c3663df5fb3bc98b3cb4a3d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f07438e20bdf71c752bb661084c835341999c561bb5442d75e177223881276f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f07438e20bdf71c752bb661084c835341999c561bb5442d75e177223881276f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:58:51Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:58:43.431198 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:58:43.432227 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99450334/tls.crt::/tmp/serving-cert-99450334/tls.key\\\\\\\"\\\\nI0126 16:58:50.828320 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:58:50.830449 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:58:50.830472 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:58:50.830509 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:58:50.830515 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:58:50.834885 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:58:50.834958 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:58:50.834991 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0126 16:58:50.834905 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0126 16:58:50.835015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:58:50.835084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:58:50.835106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:58:50.835116 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0126 16:58:50.837641 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2daa3f13d37b7ae19dfca406b6b5cfdcd4f211287f7d410193f2fee36a24553\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:09Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.238420 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191ad5d41024c88c1a4fbc30f307dd6340fd55b93b00d22f62209d0e82be286f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2a4d8be409a46ffe5702797e56a80023a72a19fb7dc5c49e5f4984cedc600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:09Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.249932 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c823748751e9938f10fab08e33b5fffff5a6d15961ce028204131b7b69a56c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:09Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.261727 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:09Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.271442 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://627cd6b39fdedd967fffcfc0755439277b2d73016fba1ddc7f5b9d5deba43b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:09Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.283814 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rq622" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a742e7b-c420-46e3-9e96-e9c744af6124\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad7222c9b91c0065a545bc1904d9864a5923cc13bfb6617daeb4a965a830f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8plh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rq622\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:09Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.295698 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63c75ede-5170-4db0-811b-5217ef8d72b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da26ddcfc6ccde3c9aabd63bdef3435f9a9eaab8c095bfeef5670c1295576cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54ca3fa13d9e8d442efa93b44a870369f7df3fe7562d77b98528f5c19a751f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xm9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:09Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.295960 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a77e85f9-b566-4807-bb92-55963c97b93c-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-v7579\" (UID: \"a77e85f9-b566-4807-bb92-55963c97b93c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v7579" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.296009 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4n9h7\" (UniqueName: \"kubernetes.io/projected/a77e85f9-b566-4807-bb92-55963c97b93c-kube-api-access-4n9h7\") pod \"ovnkube-control-plane-749d76644c-v7579\" (UID: \"a77e85f9-b566-4807-bb92-55963c97b93c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v7579" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.296100 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a77e85f9-b566-4807-bb92-55963c97b93c-env-overrides\") pod \"ovnkube-control-plane-749d76644c-v7579\" (UID: \"a77e85f9-b566-4807-bb92-55963c97b93c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v7579" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.296416 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a77e85f9-b566-4807-bb92-55963c97b93c-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-v7579\" (UID: \"a77e85f9-b566-4807-bb92-55963c97b93c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v7579" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.302236 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.302276 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.302285 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.302300 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.302311 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:09Z","lastTransitionTime":"2026-01-26T16:59:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.308302 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af4930ff9cebd62106545fb3da2dfc93f7b591426fe4c85aa2e637b60c935f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:09Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.330544 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c09748c0963bdb50c9552f249d1135ea53e68c52babd788dad050951cb849cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:09Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.344192 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:09Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.356243 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v7579" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a77e85f9-b566-4807-bb92-55963c97b93c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v7579\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:09Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.382239 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rq622" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a742e7b-c420-46e3-9e96-e9c744af6124\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad7222c9b91c0065a545bc1904d9864a5923cc13bfb6617daeb4a965a830f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8plh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rq622\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:09Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.396992 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a77e85f9-b566-4807-bb92-55963c97b93c-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-v7579\" (UID: \"a77e85f9-b566-4807-bb92-55963c97b93c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v7579" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.397040 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a77e85f9-b566-4807-bb92-55963c97b93c-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-v7579\" (UID: \"a77e85f9-b566-4807-bb92-55963c97b93c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v7579" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.397061 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4n9h7\" (UniqueName: \"kubernetes.io/projected/a77e85f9-b566-4807-bb92-55963c97b93c-kube-api-access-4n9h7\") pod \"ovnkube-control-plane-749d76644c-v7579\" (UID: \"a77e85f9-b566-4807-bb92-55963c97b93c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v7579" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.397100 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a77e85f9-b566-4807-bb92-55963c97b93c-env-overrides\") pod \"ovnkube-control-plane-749d76644c-v7579\" (UID: \"a77e85f9-b566-4807-bb92-55963c97b93c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v7579" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.397703 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a77e85f9-b566-4807-bb92-55963c97b93c-env-overrides\") pod \"ovnkube-control-plane-749d76644c-v7579\" (UID: \"a77e85f9-b566-4807-bb92-55963c97b93c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v7579" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.397868 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a77e85f9-b566-4807-bb92-55963c97b93c-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-v7579\" (UID: \"a77e85f9-b566-4807-bb92-55963c97b93c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v7579" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.402227 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ecd87a-c5db-446d-ad3e-cfabbd648c1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89975b8f9428f81ab5d3fb48ced5dd9c837bea2feea3b89f5f7ff8d7d5d15b3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d432159a07ba20bf95f058bf8a597d67cbd8d852519bed035694f8ba3d8ec4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59ba0f131eee14df389391e46bc772894e49e976c3663df5fb3bc98b3cb4a3d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03a01b651d9f66da4dd1f6e0d29ad97c0d6ae46b644c3d997d8dc99476706df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f07438e20bdf71c752bb661084c835341999c561bb5442d75e177223881276f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:58:51Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:58:43.431198 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:58:43.432227 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99450334/tls.crt::/tmp/serving-cert-99450334/tls.key\\\\\\\"\\\\nI0126 16:58:50.828320 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:58:50.830449 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:58:50.830472 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:58:50.830509 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:58:50.830515 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:58:50.834885 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:58:50.834958 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:58:50.834991 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0126 16:58:50.834905 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0126 16:58:50.835015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:58:50.835084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:58:50.835106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:58:50.835116 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0126 16:58:50.837641 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2daa3f13d37b7ae19dfca406b6b5cfdcd4f211287f7d410193f2fee36a24553\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:09Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.405116 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.405178 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.405201 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.405231 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.405253 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:09Z","lastTransitionTime":"2026-01-26T16:59:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.409098 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a77e85f9-b566-4807-bb92-55963c97b93c-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-v7579\" (UID: \"a77e85f9-b566-4807-bb92-55963c97b93c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v7579" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.428348 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4n9h7\" (UniqueName: \"kubernetes.io/projected/a77e85f9-b566-4807-bb92-55963c97b93c-kube-api-access-4n9h7\") pod \"ovnkube-control-plane-749d76644c-v7579\" (UID: \"a77e85f9-b566-4807-bb92-55963c97b93c\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v7579" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.431191 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191ad5d41024c88c1a4fbc30f307dd6340fd55b93b00d22f62209d0e82be286f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2a4d8be409a46ffe5702797e56a80023a72a19fb7dc5c49e5f4984cedc600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:09Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.444258 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c823748751e9938f10fab08e33b5fffff5a6d15961ce028204131b7b69a56c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:09Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.456287 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:09Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.465789 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 17:27:52.512562162 +0000 UTC Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.467986 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://627cd6b39fdedd967fffcfc0755439277b2d73016fba1ddc7f5b9d5deba43b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:09Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.481651 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:09Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.497995 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad7b59f9-beb7-49d6-a2d1-e29133e46854\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53a6a6d4222ee5e1ac29fd5957d8cdf3fd42de72c68c85329374ce7afc4004\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v2l7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:09Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.509277 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.509323 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.509338 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.509358 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.509371 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:09Z","lastTransitionTime":"2026-01-26T16:59:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.509696 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v7579" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.521805 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36e1e2c0e79e963e1ad9b28e9b6e7d69c6b6df040359eef2630b4aeb32109f38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pxh94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:09Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.612059 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.612099 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.612107 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.612123 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.612133 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:09Z","lastTransitionTime":"2026-01-26T16:59:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.715045 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.715081 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.715091 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.715109 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.715120 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:09Z","lastTransitionTime":"2026-01-26T16:59:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.817905 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.817944 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.817952 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.817968 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.817978 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:09Z","lastTransitionTime":"2026-01-26T16:59:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.920774 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.920813 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.920822 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.920856 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:09 crc kubenswrapper[4856]: I0126 16:59:09.920865 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:09Z","lastTransitionTime":"2026-01-26T16:59:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.023017 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.023045 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.023054 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.023068 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.023077 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:10Z","lastTransitionTime":"2026-01-26T16:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.150822 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.150867 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.150877 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.150897 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.150911 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:10Z","lastTransitionTime":"2026-01-26T16:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.155127 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v7579" event={"ID":"a77e85f9-b566-4807-bb92-55963c97b93c","Type":"ContainerStarted","Data":"8ba87c9fc35c230bbee201a5176cb467309f0b9aee82dfc81f3b677a15486d02"} Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.155165 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v7579" event={"ID":"a77e85f9-b566-4807-bb92-55963c97b93c","Type":"ContainerStarted","Data":"8f4c467dd37bd5f0ee2c1948583c7bc17be187850c547ffd809acbda9b7dd364"} Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.255983 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.256015 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.256023 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.256039 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.256051 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:10Z","lastTransitionTime":"2026-01-26T16:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.296897 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-295wr"] Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.297374 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 16:59:10 crc kubenswrapper[4856]: E0126 16:59:10.297444 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.316177 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af4930ff9cebd62106545fb3da2dfc93f7b591426fe4c85aa2e637b60c935f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:10Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.381456 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.381502 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.381515 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.381555 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.381571 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:10Z","lastTransitionTime":"2026-01-26T16:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.382272 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c09748c0963bdb50c9552f249d1135ea53e68c52babd788dad050951cb849cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:10Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.394788 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.394892 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.394796 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:59:10 crc kubenswrapper[4856]: E0126 16:59:10.394930 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:59:10 crc kubenswrapper[4856]: E0126 16:59:10.395058 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:59:10 crc kubenswrapper[4856]: E0126 16:59:10.395138 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.396547 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:10Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.408786 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v7579" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a77e85f9-b566-4807-bb92-55963c97b93c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v7579\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:10Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.410012 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tf98h\" (UniqueName: \"kubernetes.io/projected/12e50462-28e6-4531-ada4-e652310e6cce-kube-api-access-tf98h\") pod \"network-metrics-daemon-295wr\" (UID: \"12e50462-28e6-4531-ada4-e652310e6cce\") " pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.410418 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12e50462-28e6-4531-ada4-e652310e6cce-metrics-certs\") pod \"network-metrics-daemon-295wr\" (UID: \"12e50462-28e6-4531-ada4-e652310e6cce\") " pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.424696 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ecd87a-c5db-446d-ad3e-cfabbd648c1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89975b8f9428f81ab5d3fb48ced5dd9c837bea2feea3b89f5f7ff8d7d5d15b3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d432159a07ba20bf95f058bf8a597d67cbd8d852519bed035694f8ba3d8ec4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59ba0f131eee14df389391e46bc772894e49e976c3663df5fb3bc98b3cb4a3d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03a01b651d9f66da4dd1f6e0d29ad97c0d6ae46b644c3d997d8dc99476706df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f07438e20bdf71c752bb661084c835341999c561bb5442d75e177223881276f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:58:51Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:58:43.431198 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:58:43.432227 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99450334/tls.crt::/tmp/serving-cert-99450334/tls.key\\\\\\\"\\\\nI0126 16:58:50.828320 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:58:50.830449 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:58:50.830472 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:58:50.830509 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:58:50.830515 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:58:50.834885 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:58:50.834958 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:58:50.834991 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0126 16:58:50.834905 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0126 16:58:50.835015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:58:50.835084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:58:50.835106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:58:50.835116 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0126 16:58:50.837641 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2daa3f13d37b7ae19dfca406b6b5cfdcd4f211287f7d410193f2fee36a24553\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:10Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.436759 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191ad5d41024c88c1a4fbc30f307dd6340fd55b93b00d22f62209d0e82be286f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2a4d8be409a46ffe5702797e56a80023a72a19fb7dc5c49e5f4984cedc600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:10Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.448228 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c823748751e9938f10fab08e33b5fffff5a6d15961ce028204131b7b69a56c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:10Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.466848 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:10Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.467054 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 15:12:27.601921122 +0000 UTC Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.479767 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://627cd6b39fdedd967fffcfc0755439277b2d73016fba1ddc7f5b9d5deba43b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:10Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.483944 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.483989 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.484001 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.484019 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.484033 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:10Z","lastTransitionTime":"2026-01-26T16:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.493730 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rq622" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a742e7b-c420-46e3-9e96-e9c744af6124\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad7222c9b91c0065a545bc1904d9864a5923cc13bfb6617daeb4a965a830f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8plh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rq622\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:10Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.543262 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.543398 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tf98h\" (UniqueName: \"kubernetes.io/projected/12e50462-28e6-4531-ada4-e652310e6cce-kube-api-access-tf98h\") pod \"network-metrics-daemon-295wr\" (UID: \"12e50462-28e6-4531-ada4-e652310e6cce\") " pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 16:59:10 crc kubenswrapper[4856]: E0126 16:59:10.543427 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:59:26.543403556 +0000 UTC m=+62.496657537 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.543481 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.543568 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.543598 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.543627 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12e50462-28e6-4531-ada4-e652310e6cce-metrics-certs\") pod \"network-metrics-daemon-295wr\" (UID: \"12e50462-28e6-4531-ada4-e652310e6cce\") " pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 16:59:10 crc kubenswrapper[4856]: E0126 16:59:10.543630 4856 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 16:59:10 crc kubenswrapper[4856]: E0126 16:59:10.543672 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 16:59:26.543663653 +0000 UTC m=+62.496917634 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.543644 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:59:10 crc kubenswrapper[4856]: E0126 16:59:10.543775 4856 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 16:59:10 crc kubenswrapper[4856]: E0126 16:59:10.543788 4856 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 16:59:10 crc kubenswrapper[4856]: E0126 16:59:10.543798 4856 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:59:10 crc kubenswrapper[4856]: E0126 16:59:10.543810 4856 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 16:59:10 crc kubenswrapper[4856]: E0126 16:59:10.543840 4856 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 16:59:10 crc kubenswrapper[4856]: E0126 16:59:10.543847 4856 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 16:59:10 crc kubenswrapper[4856]: E0126 16:59:10.543858 4856 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:59:10 crc kubenswrapper[4856]: E0126 16:59:10.543817 4856 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 16:59:10 crc kubenswrapper[4856]: E0126 16:59:10.543824 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-26 16:59:26.543816017 +0000 UTC m=+62.497069988 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:59:10 crc kubenswrapper[4856]: E0126 16:59:10.543935 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 16:59:26.54392433 +0000 UTC m=+62.497178311 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 16:59:10 crc kubenswrapper[4856]: E0126 16:59:10.543948 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-26 16:59:26.543942511 +0000 UTC m=+62.497196492 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:59:10 crc kubenswrapper[4856]: E0126 16:59:10.543970 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12e50462-28e6-4531-ada4-e652310e6cce-metrics-certs podName:12e50462-28e6-4531-ada4-e652310e6cce nodeName:}" failed. No retries permitted until 2026-01-26 16:59:11.043962901 +0000 UTC m=+46.997216882 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/12e50462-28e6-4531-ada4-e652310e6cce-metrics-certs") pod "network-metrics-daemon-295wr" (UID: "12e50462-28e6-4531-ada4-e652310e6cce") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.587894 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.587937 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.587946 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.587964 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.587974 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:10Z","lastTransitionTime":"2026-01-26T16:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.595793 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:10Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.605830 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tf98h\" (UniqueName: \"kubernetes.io/projected/12e50462-28e6-4531-ada4-e652310e6cce-kube-api-access-tf98h\") pod \"network-metrics-daemon-295wr\" (UID: \"12e50462-28e6-4531-ada4-e652310e6cce\") " pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.612332 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad7b59f9-beb7-49d6-a2d1-e29133e46854\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53a6a6d4222ee5e1ac29fd5957d8cdf3fd42de72c68c85329374ce7afc4004\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v2l7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:10Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.632785 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36e1e2c0e79e963e1ad9b28e9b6e7d69c6b6df040359eef2630b4aeb32109f38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pxh94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:10Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.644694 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-295wr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12e50462-28e6-4531-ada4-e652310e6cce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-295wr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:10Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.657358 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63c75ede-5170-4db0-811b-5217ef8d72b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da26ddcfc6ccde3c9aabd63bdef3435f9a9eaab8c095bfeef5670c1295576cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54ca3fa13d9e8d442efa93b44a870369f7df3fe7562d77b98528f5c19a751f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xm9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:10Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.690386 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.690422 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.690431 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.690445 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.690455 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:10Z","lastTransitionTime":"2026-01-26T16:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.792836 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.792916 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.792932 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.792965 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.792982 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:10Z","lastTransitionTime":"2026-01-26T16:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.896169 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.896225 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.896241 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.896271 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.896291 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:10Z","lastTransitionTime":"2026-01-26T16:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.998582 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.998941 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.998955 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.998972 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:10 crc kubenswrapper[4856]: I0126 16:59:10.998985 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:10Z","lastTransitionTime":"2026-01-26T16:59:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.047418 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12e50462-28e6-4531-ada4-e652310e6cce-metrics-certs\") pod \"network-metrics-daemon-295wr\" (UID: \"12e50462-28e6-4531-ada4-e652310e6cce\") " pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 16:59:11 crc kubenswrapper[4856]: E0126 16:59:11.047582 4856 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 16:59:11 crc kubenswrapper[4856]: E0126 16:59:11.047637 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12e50462-28e6-4531-ada4-e652310e6cce-metrics-certs podName:12e50462-28e6-4531-ada4-e652310e6cce nodeName:}" failed. No retries permitted until 2026-01-26 16:59:12.047623477 +0000 UTC m=+48.000877458 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/12e50462-28e6-4531-ada4-e652310e6cce-metrics-certs") pod "network-metrics-daemon-295wr" (UID: "12e50462-28e6-4531-ada4-e652310e6cce") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.102102 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.102138 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.102150 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.102167 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.102203 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:11Z","lastTransitionTime":"2026-01-26T16:59:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.162659 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pxh94_ab5b6f50-172b-4535-a0f9-5d103bcab4e7/ovnkube-controller/0.log" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.166620 4856 generic.go:334] "Generic (PLEG): container finished" podID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerID="36e1e2c0e79e963e1ad9b28e9b6e7d69c6b6df040359eef2630b4aeb32109f38" exitCode=1 Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.166706 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" event={"ID":"ab5b6f50-172b-4535-a0f9-5d103bcab4e7","Type":"ContainerDied","Data":"36e1e2c0e79e963e1ad9b28e9b6e7d69c6b6df040359eef2630b4aeb32109f38"} Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.167676 4856 scope.go:117] "RemoveContainer" containerID="36e1e2c0e79e963e1ad9b28e9b6e7d69c6b6df040359eef2630b4aeb32109f38" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.168420 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v7579" event={"ID":"a77e85f9-b566-4807-bb92-55963c97b93c","Type":"ContainerStarted","Data":"c03dc794e9c2035f2e1983eacad3e51d76223cb1b82e2f402c73f9453e4bd2f0"} Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.187183 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:11Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.200133 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://627cd6b39fdedd967fffcfc0755439277b2d73016fba1ddc7f5b9d5deba43b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:11Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.204760 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.204790 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.204821 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.204836 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.204845 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:11Z","lastTransitionTime":"2026-01-26T16:59:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.216298 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rq622" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a742e7b-c420-46e3-9e96-e9c744af6124\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad7222c9b91c0065a545bc1904d9864a5923cc13bfb6617daeb4a965a830f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8plh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rq622\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:11Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.233078 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ecd87a-c5db-446d-ad3e-cfabbd648c1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89975b8f9428f81ab5d3fb48ced5dd9c837bea2feea3b89f5f7ff8d7d5d15b3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d432159a07ba20bf95f058bf8a597d67cbd8d852519bed035694f8ba3d8ec4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59ba0f131eee14df389391e46bc772894e49e976c3663df5fb3bc98b3cb4a3d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03a01b651d9f66da4dd1f6e0d29ad97c0d6ae46b644c3d997d8dc99476706df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f07438e20bdf71c752bb661084c835341999c561bb5442d75e177223881276f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:58:51Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:58:43.431198 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:58:43.432227 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99450334/tls.crt::/tmp/serving-cert-99450334/tls.key\\\\\\\"\\\\nI0126 16:58:50.828320 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:58:50.830449 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:58:50.830472 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:58:50.830509 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:58:50.830515 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:58:50.834885 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:58:50.834958 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:58:50.834991 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0126 16:58:50.834905 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0126 16:58:50.835015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:58:50.835084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:58:50.835106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:58:50.835116 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0126 16:58:50.837641 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2daa3f13d37b7ae19dfca406b6b5cfdcd4f211287f7d410193f2fee36a24553\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:11Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.246115 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191ad5d41024c88c1a4fbc30f307dd6340fd55b93b00d22f62209d0e82be286f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2a4d8be409a46ffe5702797e56a80023a72a19fb7dc5c49e5f4984cedc600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:11Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.259003 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c823748751e9938f10fab08e33b5fffff5a6d15961ce028204131b7b69a56c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:11Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.271340 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:11Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.284901 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad7b59f9-beb7-49d6-a2d1-e29133e46854\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53a6a6d4222ee5e1ac29fd5957d8cdf3fd42de72c68c85329374ce7afc4004\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v2l7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:11Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.301683 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36e1e2c0e79e963e1ad9b28e9b6e7d69c6b6df040359eef2630b4aeb32109f38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36e1e2c0e79e963e1ad9b28e9b6e7d69c6b6df040359eef2630b4aeb32109f38\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"ector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0126 16:59:10.414325 6042 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0126 16:59:10.414367 6042 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0126 16:59:10.414628 6042 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0126 16:59:10.414655 6042 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0126 16:59:10.414676 6042 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0126 16:59:10.414682 6042 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0126 16:59:10.414693 6042 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0126 16:59:10.414698 6042 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0126 16:59:10.414726 6042 factory.go:656] Stopping watch factory\\\\nI0126 16:59:10.414742 6042 ovnkube.go:599] Stopped ovnkube\\\\nI0126 16:59:10.414752 6042 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0126 16:59:10.414783 6042 handler.go:208] Removed *v1.Node event handler 2\\\\nI0126 16:59:10.414794 6042 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0126 16:59:10.414800 6042 handler.go:208] Removed *v1.Node event han\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pxh94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:11Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.307807 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.307848 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.307860 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.307879 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.307893 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:11Z","lastTransitionTime":"2026-01-26T16:59:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.315673 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-295wr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12e50462-28e6-4531-ada4-e652310e6cce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-295wr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:11Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.329107 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63c75ede-5170-4db0-811b-5217ef8d72b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da26ddcfc6ccde3c9aabd63bdef3435f9a9eaab8c095bfeef5670c1295576cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54ca3fa13d9e8d442efa93b44a870369f7df3fe7562d77b98528f5c19a751f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xm9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:11Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.342901 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v7579" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a77e85f9-b566-4807-bb92-55963c97b93c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v7579\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:11Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.356162 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af4930ff9cebd62106545fb3da2dfc93f7b591426fe4c85aa2e637b60c935f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:11Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.373320 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c09748c0963bdb50c9552f249d1135ea53e68c52babd788dad050951cb849cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:11Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.386175 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:11Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.397397 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:11Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.410860 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.410945 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.410958 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.410978 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.410990 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:11Z","lastTransitionTime":"2026-01-26T16:59:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.411447 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v7579" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a77e85f9-b566-4807-bb92-55963c97b93c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba87c9fc35c230bbee201a5176cb467309f0b9aee82dfc81f3b677a15486d02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c03dc794e9c2035f2e1983eacad3e51d76223cb1b82e2f402c73f9453e4bd2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v7579\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:11Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.424862 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af4930ff9cebd62106545fb3da2dfc93f7b591426fe4c85aa2e637b60c935f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:11Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.439972 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c09748c0963bdb50c9552f249d1135ea53e68c52babd788dad050951cb849cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:11Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.454103 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c823748751e9938f10fab08e33b5fffff5a6d15961ce028204131b7b69a56c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:11Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.467344 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 09:48:19.434967119 +0000 UTC Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.468439 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:11Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.481744 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://627cd6b39fdedd967fffcfc0755439277b2d73016fba1ddc7f5b9d5deba43b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:11Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.499113 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rq622" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a742e7b-c420-46e3-9e96-e9c744af6124\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad7222c9b91c0065a545bc1904d9864a5923cc13bfb6617daeb4a965a830f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8plh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rq622\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:11Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.513386 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.513423 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.513435 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.513451 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.513462 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:11Z","lastTransitionTime":"2026-01-26T16:59:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.513898 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ecd87a-c5db-446d-ad3e-cfabbd648c1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89975b8f9428f81ab5d3fb48ced5dd9c837bea2feea3b89f5f7ff8d7d5d15b3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d432159a07ba20bf95f058bf8a597d67cbd8d852519bed035694f8ba3d8ec4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59ba0f131eee14df389391e46bc772894e49e976c3663df5fb3bc98b3cb4a3d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03a01b651d9f66da4dd1f6e0d29ad97c0d6ae46b644c3d997d8dc99476706df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f07438e20bdf71c752bb661084c835341999c561bb5442d75e177223881276f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:58:51Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:58:43.431198 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:58:43.432227 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99450334/tls.crt::/tmp/serving-cert-99450334/tls.key\\\\\\\"\\\\nI0126 16:58:50.828320 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:58:50.830449 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:58:50.830472 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:58:50.830509 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:58:50.830515 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:58:50.834885 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:58:50.834958 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:58:50.834991 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0126 16:58:50.834905 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0126 16:58:50.835015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:58:50.835084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:58:50.835106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:58:50.835116 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0126 16:58:50.837641 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2daa3f13d37b7ae19dfca406b6b5cfdcd4f211287f7d410193f2fee36a24553\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:11Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.532961 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191ad5d41024c88c1a4fbc30f307dd6340fd55b93b00d22f62209d0e82be286f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2a4d8be409a46ffe5702797e56a80023a72a19fb7dc5c49e5f4984cedc600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:11Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.546566 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-295wr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12e50462-28e6-4531-ada4-e652310e6cce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-295wr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:11Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.568378 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:11Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.582305 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad7b59f9-beb7-49d6-a2d1-e29133e46854\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53a6a6d4222ee5e1ac29fd5957d8cdf3fd42de72c68c85329374ce7afc4004\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v2l7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:11Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.604700 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://36e1e2c0e79e963e1ad9b28e9b6e7d69c6b6df040359eef2630b4aeb32109f38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36e1e2c0e79e963e1ad9b28e9b6e7d69c6b6df040359eef2630b4aeb32109f38\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"ector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0126 16:59:10.414325 6042 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0126 16:59:10.414367 6042 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0126 16:59:10.414628 6042 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0126 16:59:10.414655 6042 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0126 16:59:10.414676 6042 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0126 16:59:10.414682 6042 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0126 16:59:10.414693 6042 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0126 16:59:10.414698 6042 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0126 16:59:10.414726 6042 factory.go:656] Stopping watch factory\\\\nI0126 16:59:10.414742 6042 ovnkube.go:599] Stopped ovnkube\\\\nI0126 16:59:10.414752 6042 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0126 16:59:10.414783 6042 handler.go:208] Removed *v1.Node event handler 2\\\\nI0126 16:59:10.414794 6042 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0126 16:59:10.414800 6042 handler.go:208] Removed *v1.Node event han\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pxh94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:11Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.615966 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.616010 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.616021 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.616041 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.616052 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:11Z","lastTransitionTime":"2026-01-26T16:59:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.617516 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63c75ede-5170-4db0-811b-5217ef8d72b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da26ddcfc6ccde3c9aabd63bdef3435f9a9eaab8c095bfeef5670c1295576cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54ca3fa13d9e8d442efa93b44a870369f7df3fe7562d77b98528f5c19a751f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xm9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:11Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.719381 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.719428 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.719438 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.719454 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.719466 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:11Z","lastTransitionTime":"2026-01-26T16:59:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.956097 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.956184 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.956224 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.956254 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:11 crc kubenswrapper[4856]: I0126 16:59:11.956271 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:11Z","lastTransitionTime":"2026-01-26T16:59:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.057352 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12e50462-28e6-4531-ada4-e652310e6cce-metrics-certs\") pod \"network-metrics-daemon-295wr\" (UID: \"12e50462-28e6-4531-ada4-e652310e6cce\") " pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 16:59:12 crc kubenswrapper[4856]: E0126 16:59:12.057808 4856 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 16:59:12 crc kubenswrapper[4856]: E0126 16:59:12.058003 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12e50462-28e6-4531-ada4-e652310e6cce-metrics-certs podName:12e50462-28e6-4531-ada4-e652310e6cce nodeName:}" failed. No retries permitted until 2026-01-26 16:59:14.05791474 +0000 UTC m=+50.011168751 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/12e50462-28e6-4531-ada4-e652310e6cce-metrics-certs") pod "network-metrics-daemon-295wr" (UID: "12e50462-28e6-4531-ada4-e652310e6cce") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.059605 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.059664 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.059677 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.059694 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.059707 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:12Z","lastTransitionTime":"2026-01-26T16:59:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.162866 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.162925 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.162936 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.162993 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.163006 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:12Z","lastTransitionTime":"2026-01-26T16:59:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.176800 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pxh94_ab5b6f50-172b-4535-a0f9-5d103bcab4e7/ovnkube-controller/0.log" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.183491 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" event={"ID":"ab5b6f50-172b-4535-a0f9-5d103bcab4e7","Type":"ContainerStarted","Data":"7c166114ee4e41f0a7e4b0590da090e98c319ef6eda0b9611419dfc55ceb139c"} Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.184297 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.199520 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:12Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.217550 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad7b59f9-beb7-49d6-a2d1-e29133e46854\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53a6a6d4222ee5e1ac29fd5957d8cdf3fd42de72c68c85329374ce7afc4004\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v2l7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:12Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.243595 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c166114ee4e41f0a7e4b0590da090e98c319ef6eda0b9611419dfc55ceb139c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36e1e2c0e79e963e1ad9b28e9b6e7d69c6b6df040359eef2630b4aeb32109f38\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"ector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0126 16:59:10.414325 6042 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0126 16:59:10.414367 6042 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0126 16:59:10.414628 6042 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0126 16:59:10.414655 6042 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0126 16:59:10.414676 6042 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0126 16:59:10.414682 6042 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0126 16:59:10.414693 6042 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0126 16:59:10.414698 6042 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0126 16:59:10.414726 6042 factory.go:656] Stopping watch factory\\\\nI0126 16:59:10.414742 6042 ovnkube.go:599] Stopped ovnkube\\\\nI0126 16:59:10.414752 6042 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0126 16:59:10.414783 6042 handler.go:208] Removed *v1.Node event handler 2\\\\nI0126 16:59:10.414794 6042 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0126 16:59:10.414800 6042 handler.go:208] Removed *v1.Node event han\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pxh94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:12Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.265813 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.265866 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.265885 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.265905 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.265918 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:12Z","lastTransitionTime":"2026-01-26T16:59:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.282519 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-295wr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12e50462-28e6-4531-ada4-e652310e6cce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-295wr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:12Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.303447 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63c75ede-5170-4db0-811b-5217ef8d72b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da26ddcfc6ccde3c9aabd63bdef3435f9a9eaab8c095bfeef5670c1295576cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54ca3fa13d9e8d442efa93b44a870369f7df3fe7562d77b98528f5c19a751f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xm9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:12Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.313497 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v7579" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a77e85f9-b566-4807-bb92-55963c97b93c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba87c9fc35c230bbee201a5176cb467309f0b9aee82dfc81f3b677a15486d02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c03dc794e9c2035f2e1983eacad3e51d76223cb1b82e2f402c73f9453e4bd2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v7579\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:12Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.325824 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af4930ff9cebd62106545fb3da2dfc93f7b591426fe4c85aa2e637b60c935f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:12Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.340393 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c09748c0963bdb50c9552f249d1135ea53e68c52babd788dad050951cb849cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:12Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.353136 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:12Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.363620 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:12Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.367727 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.367755 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.367763 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.367778 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.367788 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:12Z","lastTransitionTime":"2026-01-26T16:59:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.375356 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://627cd6b39fdedd967fffcfc0755439277b2d73016fba1ddc7f5b9d5deba43b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:12Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.388807 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rq622" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a742e7b-c420-46e3-9e96-e9c744af6124\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad7222c9b91c0065a545bc1904d9864a5923cc13bfb6617daeb4a965a830f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8plh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rq622\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:12Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.394710 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.394823 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.394799 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.394782 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:59:12 crc kubenswrapper[4856]: E0126 16:59:12.395045 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:59:12 crc kubenswrapper[4856]: E0126 16:59:12.398730 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:59:12 crc kubenswrapper[4856]: E0126 16:59:12.398919 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:59:12 crc kubenswrapper[4856]: E0126 16:59:12.399052 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.406642 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ecd87a-c5db-446d-ad3e-cfabbd648c1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89975b8f9428f81ab5d3fb48ced5dd9c837bea2feea3b89f5f7ff8d7d5d15b3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d432159a07ba20bf95f058bf8a597d67cbd8d852519bed035694f8ba3d8ec4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59ba0f131eee14df389391e46bc772894e49e976c3663df5fb3bc98b3cb4a3d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03a01b651d9f66da4dd1f6e0d29ad97c0d6ae46b644c3d997d8dc99476706df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f07438e20bdf71c752bb661084c835341999c561bb5442d75e177223881276f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:58:51Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:58:43.431198 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:58:43.432227 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99450334/tls.crt::/tmp/serving-cert-99450334/tls.key\\\\\\\"\\\\nI0126 16:58:50.828320 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:58:50.830449 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:58:50.830472 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:58:50.830509 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:58:50.830515 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:58:50.834885 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:58:50.834958 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:58:50.834991 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0126 16:58:50.834905 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0126 16:58:50.835015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:58:50.835084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:58:50.835106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:58:50.835116 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0126 16:58:50.837641 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2daa3f13d37b7ae19dfca406b6b5cfdcd4f211287f7d410193f2fee36a24553\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:12Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.422136 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191ad5d41024c88c1a4fbc30f307dd6340fd55b93b00d22f62209d0e82be286f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2a4d8be409a46ffe5702797e56a80023a72a19fb7dc5c49e5f4984cedc600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:12Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.436682 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c823748751e9938f10fab08e33b5fffff5a6d15961ce028204131b7b69a56c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:12Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.468609 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 09:57:19.547570648 +0000 UTC Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.471347 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.471388 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.471400 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.471421 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.471433 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:12Z","lastTransitionTime":"2026-01-26T16:59:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.574154 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.574305 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.574332 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.574364 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.574388 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:12Z","lastTransitionTime":"2026-01-26T16:59:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.677442 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.677493 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.677508 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.677555 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.677568 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:12Z","lastTransitionTime":"2026-01-26T16:59:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.780590 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.780640 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.780655 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.780675 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.780691 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:12Z","lastTransitionTime":"2026-01-26T16:59:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.883143 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.883194 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.883204 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.883223 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.883235 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:12Z","lastTransitionTime":"2026-01-26T16:59:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.986578 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.986651 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.986669 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.986691 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:12 crc kubenswrapper[4856]: I0126 16:59:12.986709 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:12Z","lastTransitionTime":"2026-01-26T16:59:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.088788 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.088853 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.088866 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.088882 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.088891 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:13Z","lastTransitionTime":"2026-01-26T16:59:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.189111 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pxh94_ab5b6f50-172b-4535-a0f9-5d103bcab4e7/ovnkube-controller/1.log" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.190065 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pxh94_ab5b6f50-172b-4535-a0f9-5d103bcab4e7/ovnkube-controller/0.log" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.190599 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.190664 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.190675 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.190689 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.190699 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:13Z","lastTransitionTime":"2026-01-26T16:59:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.193267 4856 generic.go:334] "Generic (PLEG): container finished" podID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerID="7c166114ee4e41f0a7e4b0590da090e98c319ef6eda0b9611419dfc55ceb139c" exitCode=1 Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.193324 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" event={"ID":"ab5b6f50-172b-4535-a0f9-5d103bcab4e7","Type":"ContainerDied","Data":"7c166114ee4e41f0a7e4b0590da090e98c319ef6eda0b9611419dfc55ceb139c"} Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.193376 4856 scope.go:117] "RemoveContainer" containerID="36e1e2c0e79e963e1ad9b28e9b6e7d69c6b6df040359eef2630b4aeb32109f38" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.194828 4856 scope.go:117] "RemoveContainer" containerID="7c166114ee4e41f0a7e4b0590da090e98c319ef6eda0b9611419dfc55ceb139c" Jan 26 16:59:13 crc kubenswrapper[4856]: E0126 16:59:13.195223 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pxh94_openshift-ovn-kubernetes(ab5b6f50-172b-4535-a0f9-5d103bcab4e7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.214944 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af4930ff9cebd62106545fb3da2dfc93f7b591426fe4c85aa2e637b60c935f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.233569 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c09748c0963bdb50c9552f249d1135ea53e68c52babd788dad050951cb849cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.245872 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.261511 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v7579" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a77e85f9-b566-4807-bb92-55963c97b93c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba87c9fc35c230bbee201a5176cb467309f0b9aee82dfc81f3b677a15486d02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c03dc794e9c2035f2e1983eacad3e51d76223cb1b82e2f402c73f9453e4bd2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v7579\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.275636 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://627cd6b39fdedd967fffcfc0755439277b2d73016fba1ddc7f5b9d5deba43b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.290287 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rq622" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a742e7b-c420-46e3-9e96-e9c744af6124\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad7222c9b91c0065a545bc1904d9864a5923cc13bfb6617daeb4a965a830f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8plh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rq622\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.293781 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.293811 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.293820 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.293833 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.293843 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:13Z","lastTransitionTime":"2026-01-26T16:59:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.304969 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ecd87a-c5db-446d-ad3e-cfabbd648c1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89975b8f9428f81ab5d3fb48ced5dd9c837bea2feea3b89f5f7ff8d7d5d15b3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d432159a07ba20bf95f058bf8a597d67cbd8d852519bed035694f8ba3d8ec4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59ba0f131eee14df389391e46bc772894e49e976c3663df5fb3bc98b3cb4a3d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03a01b651d9f66da4dd1f6e0d29ad97c0d6ae46b644c3d997d8dc99476706df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f07438e20bdf71c752bb661084c835341999c561bb5442d75e177223881276f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:58:51Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:58:43.431198 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:58:43.432227 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99450334/tls.crt::/tmp/serving-cert-99450334/tls.key\\\\\\\"\\\\nI0126 16:58:50.828320 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:58:50.830449 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:58:50.830472 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:58:50.830509 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:58:50.830515 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:58:50.834885 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:58:50.834958 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:58:50.834991 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0126 16:58:50.834905 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0126 16:58:50.835015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:58:50.835084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:58:50.835106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:58:50.835116 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0126 16:58:50.837641 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2daa3f13d37b7ae19dfca406b6b5cfdcd4f211287f7d410193f2fee36a24553\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.317429 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191ad5d41024c88c1a4fbc30f307dd6340fd55b93b00d22f62209d0e82be286f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2a4d8be409a46ffe5702797e56a80023a72a19fb7dc5c49e5f4984cedc600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.327946 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c823748751e9938f10fab08e33b5fffff5a6d15961ce028204131b7b69a56c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.339263 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.351166 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.364624 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad7b59f9-beb7-49d6-a2d1-e29133e46854\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53a6a6d4222ee5e1ac29fd5957d8cdf3fd42de72c68c85329374ce7afc4004\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v2l7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.384810 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c166114ee4e41f0a7e4b0590da090e98c319ef6eda0b9611419dfc55ceb139c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36e1e2c0e79e963e1ad9b28e9b6e7d69c6b6df040359eef2630b4aeb32109f38\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"ector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0126 16:59:10.414325 6042 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0126 16:59:10.414367 6042 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0126 16:59:10.414628 6042 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0126 16:59:10.414655 6042 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0126 16:59:10.414676 6042 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0126 16:59:10.414682 6042 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0126 16:59:10.414693 6042 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0126 16:59:10.414698 6042 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0126 16:59:10.414726 6042 factory.go:656] Stopping watch factory\\\\nI0126 16:59:10.414742 6042 ovnkube.go:599] Stopped ovnkube\\\\nI0126 16:59:10.414752 6042 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0126 16:59:10.414783 6042 handler.go:208] Removed *v1.Node event handler 2\\\\nI0126 16:59:10.414794 6042 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0126 16:59:10.414800 6042 handler.go:208] Removed *v1.Node event han\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c166114ee4e41f0a7e4b0590da090e98c319ef6eda0b9611419dfc55ceb139c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:59:12Z\\\",\\\"message\\\":\\\"etworkPolicy event handler 4 for removal\\\\nI0126 16:59:12.299079 6308 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0126 16:59:12.299145 6308 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0126 16:59:12.299187 6308 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0126 16:59:12.299270 6308 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0126 16:59:12.299298 6308 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0126 16:59:12.299298 6308 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0126 16:59:12.299306 6308 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0126 16:59:12.299340 6308 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0126 16:59:12.299299 6308 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0126 16:59:12.299385 6308 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0126 16:59:12.312081 6308 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0126 16:59:12.312141 6308 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0126 16:59:12.312158 6308 factory.go:656] Stopping watch factory\\\\nI0126 16:59:12.312172 6308 ovnkube.go:599] Stopped ovnkube\\\\nI0126 16:59:12.312201 6308 handler.go:208] Removed *v1.Node event handler 2\\\\nI01\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pxh94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.396694 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.396745 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.396756 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.396771 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.396782 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:13Z","lastTransitionTime":"2026-01-26T16:59:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.398923 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-295wr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12e50462-28e6-4531-ada4-e652310e6cce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-295wr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.415655 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63c75ede-5170-4db0-811b-5217ef8d72b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da26ddcfc6ccde3c9aabd63bdef3435f9a9eaab8c095bfeef5670c1295576cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54ca3fa13d9e8d442efa93b44a870369f7df3fe7562d77b98528f5c19a751f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xm9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.469224 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 06:14:20.520559591 +0000 UTC Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.499188 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.499224 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.499234 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.499250 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.499263 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:13Z","lastTransitionTime":"2026-01-26T16:59:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.601309 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.601348 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.601361 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.601378 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.601389 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:13Z","lastTransitionTime":"2026-01-26T16:59:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.704911 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.704944 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.704954 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.704967 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.704978 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:13Z","lastTransitionTime":"2026-01-26T16:59:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.720196 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.777094 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af4930ff9cebd62106545fb3da2dfc93f7b591426fe4c85aa2e637b60c935f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.779388 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.791868 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c09748c0963bdb50c9552f249d1135ea53e68c52babd788dad050951cb849cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.803445 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.807044 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.807099 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.807112 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.807128 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.807138 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:13Z","lastTransitionTime":"2026-01-26T16:59:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.817008 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v7579" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a77e85f9-b566-4807-bb92-55963c97b93c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba87c9fc35c230bbee201a5176cb467309f0b9aee82dfc81f3b677a15486d02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c03dc794e9c2035f2e1983eacad3e51d76223cb1b82e2f402c73f9453e4bd2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v7579\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.830049 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ecd87a-c5db-446d-ad3e-cfabbd648c1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89975b8f9428f81ab5d3fb48ced5dd9c837bea2feea3b89f5f7ff8d7d5d15b3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d432159a07ba20bf95f058bf8a597d67cbd8d852519bed035694f8ba3d8ec4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59ba0f131eee14df389391e46bc772894e49e976c3663df5fb3bc98b3cb4a3d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03a01b651d9f66da4dd1f6e0d29ad97c0d6ae46b644c3d997d8dc99476706df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f07438e20bdf71c752bb661084c835341999c561bb5442d75e177223881276f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:58:51Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:58:43.431198 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:58:43.432227 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99450334/tls.crt::/tmp/serving-cert-99450334/tls.key\\\\\\\"\\\\nI0126 16:58:50.828320 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:58:50.830449 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:58:50.830472 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:58:50.830509 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:58:50.830515 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:58:50.834885 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:58:50.834958 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:58:50.834991 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0126 16:58:50.834905 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0126 16:58:50.835015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:58:50.835084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:58:50.835106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:58:50.835116 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0126 16:58:50.837641 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2daa3f13d37b7ae19dfca406b6b5cfdcd4f211287f7d410193f2fee36a24553\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.847966 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191ad5d41024c88c1a4fbc30f307dd6340fd55b93b00d22f62209d0e82be286f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2a4d8be409a46ffe5702797e56a80023a72a19fb7dc5c49e5f4984cedc600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.859969 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c823748751e9938f10fab08e33b5fffff5a6d15961ce028204131b7b69a56c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.871947 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.882058 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://627cd6b39fdedd967fffcfc0755439277b2d73016fba1ddc7f5b9d5deba43b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.894242 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rq622" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a742e7b-c420-46e3-9e96-e9c744af6124\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad7222c9b91c0065a545bc1904d9864a5923cc13bfb6617daeb4a965a830f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8plh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rq622\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.907511 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.908887 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.908944 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.908956 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.908977 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.908990 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:13Z","lastTransitionTime":"2026-01-26T16:59:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.924582 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad7b59f9-beb7-49d6-a2d1-e29133e46854\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53a6a6d4222ee5e1ac29fd5957d8cdf3fd42de72c68c85329374ce7afc4004\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v2l7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.942418 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c166114ee4e41f0a7e4b0590da090e98c319ef6eda0b9611419dfc55ceb139c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36e1e2c0e79e963e1ad9b28e9b6e7d69c6b6df040359eef2630b4aeb32109f38\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"ector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0126 16:59:10.414325 6042 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0126 16:59:10.414367 6042 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0126 16:59:10.414628 6042 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0126 16:59:10.414655 6042 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0126 16:59:10.414676 6042 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0126 16:59:10.414682 6042 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0126 16:59:10.414693 6042 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0126 16:59:10.414698 6042 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0126 16:59:10.414726 6042 factory.go:656] Stopping watch factory\\\\nI0126 16:59:10.414742 6042 ovnkube.go:599] Stopped ovnkube\\\\nI0126 16:59:10.414752 6042 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0126 16:59:10.414783 6042 handler.go:208] Removed *v1.Node event handler 2\\\\nI0126 16:59:10.414794 6042 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0126 16:59:10.414800 6042 handler.go:208] Removed *v1.Node event han\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:03Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c166114ee4e41f0a7e4b0590da090e98c319ef6eda0b9611419dfc55ceb139c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:59:12Z\\\",\\\"message\\\":\\\"etworkPolicy event handler 4 for removal\\\\nI0126 16:59:12.299079 6308 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0126 16:59:12.299145 6308 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0126 16:59:12.299187 6308 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0126 16:59:12.299270 6308 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0126 16:59:12.299298 6308 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0126 16:59:12.299298 6308 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0126 16:59:12.299306 6308 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0126 16:59:12.299340 6308 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0126 16:59:12.299299 6308 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0126 16:59:12.299385 6308 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0126 16:59:12.312081 6308 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0126 16:59:12.312141 6308 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0126 16:59:12.312158 6308 factory.go:656] Stopping watch factory\\\\nI0126 16:59:12.312172 6308 ovnkube.go:599] Stopped ovnkube\\\\nI0126 16:59:12.312201 6308 handler.go:208] Removed *v1.Node event handler 2\\\\nI01\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pxh94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.953859 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-295wr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12e50462-28e6-4531-ada4-e652310e6cce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-295wr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:13 crc kubenswrapper[4856]: I0126 16:59:13.969060 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63c75ede-5170-4db0-811b-5217ef8d72b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da26ddcfc6ccde3c9aabd63bdef3435f9a9eaab8c095bfeef5670c1295576cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54ca3fa13d9e8d442efa93b44a870369f7df3fe7562d77b98528f5c19a751f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xm9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:13Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.011191 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.011231 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.011252 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.011267 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.011275 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:14Z","lastTransitionTime":"2026-01-26T16:59:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.079274 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12e50462-28e6-4531-ada4-e652310e6cce-metrics-certs\") pod \"network-metrics-daemon-295wr\" (UID: \"12e50462-28e6-4531-ada4-e652310e6cce\") " pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 16:59:14 crc kubenswrapper[4856]: E0126 16:59:14.079482 4856 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 16:59:14 crc kubenswrapper[4856]: E0126 16:59:14.079604 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12e50462-28e6-4531-ada4-e652310e6cce-metrics-certs podName:12e50462-28e6-4531-ada4-e652310e6cce nodeName:}" failed. No retries permitted until 2026-01-26 16:59:18.079584776 +0000 UTC m=+54.032838757 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/12e50462-28e6-4531-ada4-e652310e6cce-metrics-certs") pod "network-metrics-daemon-295wr" (UID: "12e50462-28e6-4531-ada4-e652310e6cce") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.113973 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.114012 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.114024 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.114056 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.114066 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:14Z","lastTransitionTime":"2026-01-26T16:59:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.197841 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pxh94_ab5b6f50-172b-4535-a0f9-5d103bcab4e7/ovnkube-controller/1.log" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.201204 4856 scope.go:117] "RemoveContainer" containerID="7c166114ee4e41f0a7e4b0590da090e98c319ef6eda0b9611419dfc55ceb139c" Jan 26 16:59:14 crc kubenswrapper[4856]: E0126 16:59:14.201369 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pxh94_openshift-ovn-kubernetes(ab5b6f50-172b-4535-a0f9-5d103bcab4e7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.212785 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc864b0d-83bc-4954-9c61-ad650157caff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbde934a6c8acad10ca3ab8206d0ddbd4f7b17e9d304b898a68f4d3b0303bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b48f763d4aff37169399be766d5ab4f7ebbf91f304d139c9022a8556946eb107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0094e662f53c4832a984e05a880021af05ffc4c27f25394c28a070d9ef5490d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb95f623fa4f5f42217649ccce4225f9fe588bb3558da9334eef2b40ddb62486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb95f623fa4f5f42217649ccce4225f9fe588bb3558da9334eef2b40ddb62486\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:14Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.216100 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.216127 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.216135 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.216152 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.216162 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:14Z","lastTransitionTime":"2026-01-26T16:59:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.224753 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63c75ede-5170-4db0-811b-5217ef8d72b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da26ddcfc6ccde3c9aabd63bdef3435f9a9eaab8c095bfeef5670c1295576cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54ca3fa13d9e8d442efa93b44a870369f7df3fe7562d77b98528f5c19a751f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xm9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:14Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.235207 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af4930ff9cebd62106545fb3da2dfc93f7b591426fe4c85aa2e637b60c935f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:14Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.248193 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c09748c0963bdb50c9552f249d1135ea53e68c52babd788dad050951cb849cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:14Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.258909 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:14Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.269642 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v7579" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a77e85f9-b566-4807-bb92-55963c97b93c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba87c9fc35c230bbee201a5176cb467309f0b9aee82dfc81f3b677a15486d02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c03dc794e9c2035f2e1983eacad3e51d76223cb1b82e2f402c73f9453e4bd2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v7579\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:14Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.278908 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://627cd6b39fdedd967fffcfc0755439277b2d73016fba1ddc7f5b9d5deba43b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:14Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.283453 4856 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.294391 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rq622" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a742e7b-c420-46e3-9e96-e9c744af6124\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad7222c9b91c0065a545bc1904d9864a5923cc13bfb6617daeb4a965a830f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8plh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rq622\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:14Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.311471 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ecd87a-c5db-446d-ad3e-cfabbd648c1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89975b8f9428f81ab5d3fb48ced5dd9c837bea2feea3b89f5f7ff8d7d5d15b3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d432159a07ba20bf95f058bf8a597d67cbd8d852519bed035694f8ba3d8ec4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59ba0f131eee14df389391e46bc772894e49e976c3663df5fb3bc98b3cb4a3d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03a01b651d9f66da4dd1f6e0d29ad97c0d6ae46b644c3d997d8dc99476706df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f07438e20bdf71c752bb661084c835341999c561bb5442d75e177223881276f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:58:51Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:58:43.431198 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:58:43.432227 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99450334/tls.crt::/tmp/serving-cert-99450334/tls.key\\\\\\\"\\\\nI0126 16:58:50.828320 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:58:50.830449 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:58:50.830472 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:58:50.830509 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:58:50.830515 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:58:50.834885 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:58:50.834958 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:58:50.834991 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0126 16:58:50.834905 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0126 16:58:50.835015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:58:50.835084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:58:50.835106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:58:50.835116 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0126 16:58:50.837641 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2daa3f13d37b7ae19dfca406b6b5cfdcd4f211287f7d410193f2fee36a24553\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:14Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.319065 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.319118 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.319128 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.319142 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.319152 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:14Z","lastTransitionTime":"2026-01-26T16:59:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.324427 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191ad5d41024c88c1a4fbc30f307dd6340fd55b93b00d22f62209d0e82be286f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2a4d8be409a46ffe5702797e56a80023a72a19fb7dc5c49e5f4984cedc600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:14Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.338080 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c823748751e9938f10fab08e33b5fffff5a6d15961ce028204131b7b69a56c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:14Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.351285 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:14Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.362425 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:14Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.376174 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad7b59f9-beb7-49d6-a2d1-e29133e46854\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53a6a6d4222ee5e1ac29fd5957d8cdf3fd42de72c68c85329374ce7afc4004\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v2l7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:14Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.394926 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:59:14 crc kubenswrapper[4856]: E0126 16:59:14.395053 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.394936 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.394950 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:59:14 crc kubenswrapper[4856]: E0126 16:59:14.395215 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.394933 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:59:14 crc kubenswrapper[4856]: E0126 16:59:14.395313 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:59:14 crc kubenswrapper[4856]: E0126 16:59:14.395252 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.399955 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c166114ee4e41f0a7e4b0590da090e98c319ef6eda0b9611419dfc55ceb139c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c166114ee4e41f0a7e4b0590da090e98c319ef6eda0b9611419dfc55ceb139c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:59:12Z\\\",\\\"message\\\":\\\"etworkPolicy event handler 4 for removal\\\\nI0126 16:59:12.299079 6308 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0126 16:59:12.299145 6308 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0126 16:59:12.299187 6308 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0126 16:59:12.299270 6308 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0126 16:59:12.299298 6308 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0126 16:59:12.299298 6308 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0126 16:59:12.299306 6308 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0126 16:59:12.299340 6308 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0126 16:59:12.299299 6308 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0126 16:59:12.299385 6308 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0126 16:59:12.312081 6308 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0126 16:59:12.312141 6308 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0126 16:59:12.312158 6308 factory.go:656] Stopping watch factory\\\\nI0126 16:59:12.312172 6308 ovnkube.go:599] Stopped ovnkube\\\\nI0126 16:59:12.312201 6308 handler.go:208] Removed *v1.Node event handler 2\\\\nI01\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pxh94_openshift-ovn-kubernetes(ab5b6f50-172b-4535-a0f9-5d103bcab4e7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pxh94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:14Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.410638 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-295wr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12e50462-28e6-4531-ada4-e652310e6cce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-295wr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:14Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.421495 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.421535 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.421548 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.421565 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.421575 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:14Z","lastTransitionTime":"2026-01-26T16:59:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.470379 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 03:51:58.276263714 +0000 UTC Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.523888 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.523924 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.523935 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.523950 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.523960 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:14Z","lastTransitionTime":"2026-01-26T16:59:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.626379 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.626428 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.626441 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.626458 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.626474 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:14Z","lastTransitionTime":"2026-01-26T16:59:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.729008 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.729062 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.729075 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.729095 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.729108 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:14Z","lastTransitionTime":"2026-01-26T16:59:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.831574 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.831608 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.831625 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.831640 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.831650 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:14Z","lastTransitionTime":"2026-01-26T16:59:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.934096 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.934146 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.934159 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.934177 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:14 crc kubenswrapper[4856]: I0126 16:59:14.934189 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:14Z","lastTransitionTime":"2026-01-26T16:59:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.037599 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.037657 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.037673 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.037696 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.037711 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:15Z","lastTransitionTime":"2026-01-26T16:59:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.140445 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.140483 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.140492 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.140506 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.140552 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:15Z","lastTransitionTime":"2026-01-26T16:59:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.243041 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.243087 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.243099 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.243116 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.243128 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:15Z","lastTransitionTime":"2026-01-26T16:59:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.345669 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.345712 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.345727 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.345753 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.345768 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:15Z","lastTransitionTime":"2026-01-26T16:59:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.408811 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:15Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.419913 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://627cd6b39fdedd967fffcfc0755439277b2d73016fba1ddc7f5b9d5deba43b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:15Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.437160 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rq622" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a742e7b-c420-46e3-9e96-e9c744af6124\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad7222c9b91c0065a545bc1904d9864a5923cc13bfb6617daeb4a965a830f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8plh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rq622\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:15Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.449363 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.449410 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.449425 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.449448 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.449470 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:15Z","lastTransitionTime":"2026-01-26T16:59:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.453627 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ecd87a-c5db-446d-ad3e-cfabbd648c1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89975b8f9428f81ab5d3fb48ced5dd9c837bea2feea3b89f5f7ff8d7d5d15b3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d432159a07ba20bf95f058bf8a597d67cbd8d852519bed035694f8ba3d8ec4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59ba0f131eee14df389391e46bc772894e49e976c3663df5fb3bc98b3cb4a3d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03a01b651d9f66da4dd1f6e0d29ad97c0d6ae46b644c3d997d8dc99476706df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f07438e20bdf71c752bb661084c835341999c561bb5442d75e177223881276f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:58:51Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:58:43.431198 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:58:43.432227 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99450334/tls.crt::/tmp/serving-cert-99450334/tls.key\\\\\\\"\\\\nI0126 16:58:50.828320 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:58:50.830449 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:58:50.830472 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:58:50.830509 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:58:50.830515 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:58:50.834885 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:58:50.834958 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:58:50.834991 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0126 16:58:50.834905 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0126 16:58:50.835015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:58:50.835084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:58:50.835106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:58:50.835116 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0126 16:58:50.837641 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2daa3f13d37b7ae19dfca406b6b5cfdcd4f211287f7d410193f2fee36a24553\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:15Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.470302 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191ad5d41024c88c1a4fbc30f307dd6340fd55b93b00d22f62209d0e82be286f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2a4d8be409a46ffe5702797e56a80023a72a19fb7dc5c49e5f4984cedc600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:15Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.470664 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 22:45:20.21076415 +0000 UTC Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.484835 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c823748751e9938f10fab08e33b5fffff5a6d15961ce028204131b7b69a56c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:15Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.498127 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:15Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.516293 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad7b59f9-beb7-49d6-a2d1-e29133e46854\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53a6a6d4222ee5e1ac29fd5957d8cdf3fd42de72c68c85329374ce7afc4004\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v2l7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:15Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.534065 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c166114ee4e41f0a7e4b0590da090e98c319ef6eda0b9611419dfc55ceb139c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c166114ee4e41f0a7e4b0590da090e98c319ef6eda0b9611419dfc55ceb139c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:59:12Z\\\",\\\"message\\\":\\\"etworkPolicy event handler 4 for removal\\\\nI0126 16:59:12.299079 6308 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0126 16:59:12.299145 6308 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0126 16:59:12.299187 6308 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0126 16:59:12.299270 6308 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0126 16:59:12.299298 6308 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0126 16:59:12.299298 6308 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0126 16:59:12.299306 6308 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0126 16:59:12.299340 6308 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0126 16:59:12.299299 6308 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0126 16:59:12.299385 6308 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0126 16:59:12.312081 6308 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0126 16:59:12.312141 6308 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0126 16:59:12.312158 6308 factory.go:656] Stopping watch factory\\\\nI0126 16:59:12.312172 6308 ovnkube.go:599] Stopped ovnkube\\\\nI0126 16:59:12.312201 6308 handler.go:208] Removed *v1.Node event handler 2\\\\nI01\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pxh94_openshift-ovn-kubernetes(ab5b6f50-172b-4535-a0f9-5d103bcab4e7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pxh94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:15Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.546505 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-295wr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12e50462-28e6-4531-ada4-e652310e6cce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-295wr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:15Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.551639 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.551671 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.551681 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.551716 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.551731 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:15Z","lastTransitionTime":"2026-01-26T16:59:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.559709 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc864b0d-83bc-4954-9c61-ad650157caff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbde934a6c8acad10ca3ab8206d0ddbd4f7b17e9d304b898a68f4d3b0303bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b48f763d4aff37169399be766d5ab4f7ebbf91f304d139c9022a8556946eb107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0094e662f53c4832a984e05a880021af05ffc4c27f25394c28a070d9ef5490d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb95f623fa4f5f42217649ccce4225f9fe588bb3558da9334eef2b40ddb62486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb95f623fa4f5f42217649ccce4225f9fe588bb3558da9334eef2b40ddb62486\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:15Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.577017 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63c75ede-5170-4db0-811b-5217ef8d72b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da26ddcfc6ccde3c9aabd63bdef3435f9a9eaab8c095bfeef5670c1295576cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54ca3fa13d9e8d442efa93b44a870369f7df3fe7562d77b98528f5c19a751f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xm9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:15Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.587506 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v7579" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a77e85f9-b566-4807-bb92-55963c97b93c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba87c9fc35c230bbee201a5176cb467309f0b9aee82dfc81f3b677a15486d02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c03dc794e9c2035f2e1983eacad3e51d76223cb1b82e2f402c73f9453e4bd2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v7579\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:15Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.597934 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af4930ff9cebd62106545fb3da2dfc93f7b591426fe4c85aa2e637b60c935f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:15Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.610120 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c09748c0963bdb50c9552f249d1135ea53e68c52babd788dad050951cb849cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:15Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.621908 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:15Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.653857 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.653895 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.653903 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.653916 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.653925 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:15Z","lastTransitionTime":"2026-01-26T16:59:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.757097 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.757152 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.757168 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.757191 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.757207 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:15Z","lastTransitionTime":"2026-01-26T16:59:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.859461 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.859502 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.859510 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.859549 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.859562 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:15Z","lastTransitionTime":"2026-01-26T16:59:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.962381 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.962696 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.962799 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.962900 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:15 crc kubenswrapper[4856]: I0126 16:59:15.962999 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:15Z","lastTransitionTime":"2026-01-26T16:59:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.065897 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.065936 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.065944 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.065956 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.065965 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:16Z","lastTransitionTime":"2026-01-26T16:59:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.168978 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.169361 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.169715 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.169964 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.170193 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:16Z","lastTransitionTime":"2026-01-26T16:59:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.272844 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.272879 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.272892 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.272911 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.272924 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:16Z","lastTransitionTime":"2026-01-26T16:59:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.376458 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.376497 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.376509 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.376564 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.376590 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:16Z","lastTransitionTime":"2026-01-26T16:59:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.395060 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.395090 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:59:16 crc kubenswrapper[4856]: E0126 16:59:16.395181 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:59:16 crc kubenswrapper[4856]: E0126 16:59:16.395295 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.395402 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:59:16 crc kubenswrapper[4856]: E0126 16:59:16.395574 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.395697 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 16:59:16 crc kubenswrapper[4856]: E0126 16:59:16.395822 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.471043 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 07:46:13.101827915 +0000 UTC Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.479345 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.479406 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.479417 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.479443 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.479460 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:16Z","lastTransitionTime":"2026-01-26T16:59:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.583101 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.583188 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.583363 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.583433 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.583457 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:16Z","lastTransitionTime":"2026-01-26T16:59:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.686682 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.686753 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.686771 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.686796 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.686813 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:16Z","lastTransitionTime":"2026-01-26T16:59:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.790485 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.790643 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.790689 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.790734 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.790756 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:16Z","lastTransitionTime":"2026-01-26T16:59:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.893784 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.893833 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.893847 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.893873 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.893887 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:16Z","lastTransitionTime":"2026-01-26T16:59:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.996740 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.996804 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.996828 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.996860 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:16 crc kubenswrapper[4856]: I0126 16:59:16.996882 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:16Z","lastTransitionTime":"2026-01-26T16:59:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.100635 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.100693 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.100708 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.100728 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.100745 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:17Z","lastTransitionTime":"2026-01-26T16:59:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.204843 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.204910 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.205006 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.205036 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.205052 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:17Z","lastTransitionTime":"2026-01-26T16:59:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.308334 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.308393 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.308409 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.308431 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.308453 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:17Z","lastTransitionTime":"2026-01-26T16:59:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.411439 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.411483 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.411500 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.411517 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.411580 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:17Z","lastTransitionTime":"2026-01-26T16:59:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.472032 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 19:59:29.617704909 +0000 UTC Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.514569 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.514620 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.514634 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.514659 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.514674 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:17Z","lastTransitionTime":"2026-01-26T16:59:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.617181 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.617235 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.617249 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.617269 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.617283 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:17Z","lastTransitionTime":"2026-01-26T16:59:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.720505 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.720611 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.720635 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.720668 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.720691 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:17Z","lastTransitionTime":"2026-01-26T16:59:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.823409 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.823498 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.823508 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.823565 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.823580 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:17Z","lastTransitionTime":"2026-01-26T16:59:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.926430 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.926482 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.926500 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.926553 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:17 crc kubenswrapper[4856]: I0126 16:59:17.926569 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:17Z","lastTransitionTime":"2026-01-26T16:59:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.030275 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.030385 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.030398 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.030430 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.030445 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:18Z","lastTransitionTime":"2026-01-26T16:59:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.114030 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12e50462-28e6-4531-ada4-e652310e6cce-metrics-certs\") pod \"network-metrics-daemon-295wr\" (UID: \"12e50462-28e6-4531-ada4-e652310e6cce\") " pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 16:59:18 crc kubenswrapper[4856]: E0126 16:59:18.114310 4856 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 16:59:18 crc kubenswrapper[4856]: E0126 16:59:18.114414 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12e50462-28e6-4531-ada4-e652310e6cce-metrics-certs podName:12e50462-28e6-4531-ada4-e652310e6cce nodeName:}" failed. No retries permitted until 2026-01-26 16:59:26.114385653 +0000 UTC m=+62.067639684 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/12e50462-28e6-4531-ada4-e652310e6cce-metrics-certs") pod "network-metrics-daemon-295wr" (UID: "12e50462-28e6-4531-ada4-e652310e6cce") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.132908 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.132955 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.132974 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.132991 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.133002 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:18Z","lastTransitionTime":"2026-01-26T16:59:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.236616 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.236684 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.236706 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.236732 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.236748 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:18Z","lastTransitionTime":"2026-01-26T16:59:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.341476 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.341593 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.341622 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.341668 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.341700 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:18Z","lastTransitionTime":"2026-01-26T16:59:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.394352 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.394393 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:59:18 crc kubenswrapper[4856]: E0126 16:59:18.394647 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:59:18 crc kubenswrapper[4856]: E0126 16:59:18.394786 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.394399 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.395044 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:59:18 crc kubenswrapper[4856]: E0126 16:59:18.395467 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 16:59:18 crc kubenswrapper[4856]: E0126 16:59:18.396003 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.444807 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.444863 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.444875 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.444890 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.444900 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:18Z","lastTransitionTime":"2026-01-26T16:59:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.472522 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 15:18:20.671581864 +0000 UTC Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.548944 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.549012 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.549034 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.549064 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.549087 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:18Z","lastTransitionTime":"2026-01-26T16:59:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.652339 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.652409 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.652438 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.652460 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.652475 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:18Z","lastTransitionTime":"2026-01-26T16:59:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.656767 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.656839 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.656864 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.656892 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.656919 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:18Z","lastTransitionTime":"2026-01-26T16:59:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:18 crc kubenswrapper[4856]: E0126 16:59:18.677511 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17523591-a778-4a97-aeab-8a7a93101850\\\",\\\"systemUUID\\\":\\\"ca45d056-99cb-4442-8a44-7e899628ecb2\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:18Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.683713 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.683773 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.683793 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.683821 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.683846 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:18Z","lastTransitionTime":"2026-01-26T16:59:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:18 crc kubenswrapper[4856]: E0126 16:59:18.699704 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17523591-a778-4a97-aeab-8a7a93101850\\\",\\\"systemUUID\\\":\\\"ca45d056-99cb-4442-8a44-7e899628ecb2\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:18Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.704497 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.704574 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.704587 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.704606 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.704617 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:18Z","lastTransitionTime":"2026-01-26T16:59:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:18 crc kubenswrapper[4856]: E0126 16:59:18.717231 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17523591-a778-4a97-aeab-8a7a93101850\\\",\\\"systemUUID\\\":\\\"ca45d056-99cb-4442-8a44-7e899628ecb2\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:18Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.723208 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.723268 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.723283 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.723358 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.723378 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:18Z","lastTransitionTime":"2026-01-26T16:59:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:18 crc kubenswrapper[4856]: E0126 16:59:18.742506 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17523591-a778-4a97-aeab-8a7a93101850\\\",\\\"systemUUID\\\":\\\"ca45d056-99cb-4442-8a44-7e899628ecb2\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:18Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.747988 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.748266 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.748385 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.748550 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.748687 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:18Z","lastTransitionTime":"2026-01-26T16:59:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:18 crc kubenswrapper[4856]: E0126 16:59:18.765308 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17523591-a778-4a97-aeab-8a7a93101850\\\",\\\"systemUUID\\\":\\\"ca45d056-99cb-4442-8a44-7e899628ecb2\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:18Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:18 crc kubenswrapper[4856]: E0126 16:59:18.765698 4856 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.768753 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.769004 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.769160 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.769294 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.769429 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:18Z","lastTransitionTime":"2026-01-26T16:59:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.873905 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.874303 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.874482 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.874682 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.874872 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:18Z","lastTransitionTime":"2026-01-26T16:59:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.977636 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.977907 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.977993 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.978074 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:18 crc kubenswrapper[4856]: I0126 16:59:18.978191 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:18Z","lastTransitionTime":"2026-01-26T16:59:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.081355 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.081410 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.081427 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.081446 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.081456 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:19Z","lastTransitionTime":"2026-01-26T16:59:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.184613 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.184702 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.184714 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.184740 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.184753 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:19Z","lastTransitionTime":"2026-01-26T16:59:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.287814 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.287888 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.287905 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.287930 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.287944 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:19Z","lastTransitionTime":"2026-01-26T16:59:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.391467 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.391775 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.391793 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.391819 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.391843 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:19Z","lastTransitionTime":"2026-01-26T16:59:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.473759 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 01:23:04.518023633 +0000 UTC Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.495579 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.495621 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.495645 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.495663 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.495675 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:19Z","lastTransitionTime":"2026-01-26T16:59:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.597470 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.597554 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.597566 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.597582 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.597593 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:19Z","lastTransitionTime":"2026-01-26T16:59:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.700890 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.700944 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.700957 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.700983 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.700995 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:19Z","lastTransitionTime":"2026-01-26T16:59:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.803378 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.803431 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.803443 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.803464 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.803477 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:19Z","lastTransitionTime":"2026-01-26T16:59:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.906854 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.906929 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.906943 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.906965 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:19 crc kubenswrapper[4856]: I0126 16:59:19.906978 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:19Z","lastTransitionTime":"2026-01-26T16:59:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.010167 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.010239 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.010253 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.010276 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.010292 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:20Z","lastTransitionTime":"2026-01-26T16:59:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.114231 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.114312 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.114333 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.114362 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.114378 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:20Z","lastTransitionTime":"2026-01-26T16:59:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.217982 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.218028 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.218039 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.218058 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.218070 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:20Z","lastTransitionTime":"2026-01-26T16:59:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.320369 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.320416 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.320432 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.320448 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.320460 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:20Z","lastTransitionTime":"2026-01-26T16:59:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.394163 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.394163 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.394282 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.394304 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 16:59:20 crc kubenswrapper[4856]: E0126 16:59:20.394632 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:59:20 crc kubenswrapper[4856]: E0126 16:59:20.394780 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 16:59:20 crc kubenswrapper[4856]: E0126 16:59:20.394927 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:59:20 crc kubenswrapper[4856]: E0126 16:59:20.395133 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.429946 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.429998 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.430006 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.430019 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.430028 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:20Z","lastTransitionTime":"2026-01-26T16:59:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.474238 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 15:09:37.534428491 +0000 UTC Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.532490 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.532558 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.532568 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.532585 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.532594 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:20Z","lastTransitionTime":"2026-01-26T16:59:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.635249 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.635288 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.635301 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.635317 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.635327 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:20Z","lastTransitionTime":"2026-01-26T16:59:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.738642 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.738709 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.738747 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.738780 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.738888 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:20Z","lastTransitionTime":"2026-01-26T16:59:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.842355 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.842418 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.842437 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.842463 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.842481 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:20Z","lastTransitionTime":"2026-01-26T16:59:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.945242 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.945345 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.945368 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.945391 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:20 crc kubenswrapper[4856]: I0126 16:59:20.945409 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:20Z","lastTransitionTime":"2026-01-26T16:59:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.048026 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.048072 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.048092 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.048109 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.048126 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:21Z","lastTransitionTime":"2026-01-26T16:59:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.150441 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.150504 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.150515 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.150551 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.150561 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:21Z","lastTransitionTime":"2026-01-26T16:59:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.253702 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.253747 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.253763 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.254003 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.254029 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:21Z","lastTransitionTime":"2026-01-26T16:59:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.357518 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.357624 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.357647 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.357693 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.357718 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:21Z","lastTransitionTime":"2026-01-26T16:59:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.460777 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.460828 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.460844 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.460862 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.460873 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:21Z","lastTransitionTime":"2026-01-26T16:59:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.474993 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 16:29:44.323268258 +0000 UTC Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.563337 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.563367 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.563377 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.563392 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.563402 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:21Z","lastTransitionTime":"2026-01-26T16:59:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.666657 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.666713 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.666728 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.666749 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.666765 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:21Z","lastTransitionTime":"2026-01-26T16:59:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.769667 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.769706 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.769717 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.769733 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.769758 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:21Z","lastTransitionTime":"2026-01-26T16:59:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.872646 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.872771 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.872793 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.872821 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.872841 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:21Z","lastTransitionTime":"2026-01-26T16:59:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.975058 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.975124 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.975150 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.975180 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:21 crc kubenswrapper[4856]: I0126 16:59:21.975205 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:21Z","lastTransitionTime":"2026-01-26T16:59:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.078412 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.078449 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.078459 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.078488 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.078497 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:22Z","lastTransitionTime":"2026-01-26T16:59:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.180842 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.180895 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.180907 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.180926 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.180937 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:22Z","lastTransitionTime":"2026-01-26T16:59:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.282801 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.282934 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.282953 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.282971 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.282981 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:22Z","lastTransitionTime":"2026-01-26T16:59:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.385491 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.385604 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.385634 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.385664 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.385733 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:22Z","lastTransitionTime":"2026-01-26T16:59:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.394764 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.394809 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.394824 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.394771 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:59:22 crc kubenswrapper[4856]: E0126 16:59:22.394940 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:59:22 crc kubenswrapper[4856]: E0126 16:59:22.395073 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 16:59:22 crc kubenswrapper[4856]: E0126 16:59:22.395182 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:59:22 crc kubenswrapper[4856]: E0126 16:59:22.395347 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.475182 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 20:16:05.570387367 +0000 UTC Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.488321 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.488408 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.488418 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.488434 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.488451 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:22Z","lastTransitionTime":"2026-01-26T16:59:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.591768 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.591868 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.591879 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.591918 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.591929 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:22Z","lastTransitionTime":"2026-01-26T16:59:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.694761 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.694839 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.694855 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.694885 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.694902 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:22Z","lastTransitionTime":"2026-01-26T16:59:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.798495 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.798559 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.798573 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.798665 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.798683 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:22Z","lastTransitionTime":"2026-01-26T16:59:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.901577 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.901832 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.901855 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.901883 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:22 crc kubenswrapper[4856]: I0126 16:59:22.901905 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:22Z","lastTransitionTime":"2026-01-26T16:59:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.006119 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.006182 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.006198 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.006220 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.006236 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:23Z","lastTransitionTime":"2026-01-26T16:59:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.109299 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.109351 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.109359 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.109375 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.109385 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:23Z","lastTransitionTime":"2026-01-26T16:59:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.279254 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.279280 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.279288 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.279301 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.279311 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:23Z","lastTransitionTime":"2026-01-26T16:59:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.381546 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.381592 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.381600 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.381615 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.381624 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:23Z","lastTransitionTime":"2026-01-26T16:59:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.476083 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 17:31:29.554584671 +0000 UTC Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.484233 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.484267 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.484276 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.484290 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.484300 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:23Z","lastTransitionTime":"2026-01-26T16:59:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.587782 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.587820 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.587827 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.587848 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.587861 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:23Z","lastTransitionTime":"2026-01-26T16:59:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.692370 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.692431 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.692454 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.692487 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.692508 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:23Z","lastTransitionTime":"2026-01-26T16:59:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.795145 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.796063 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.796175 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.796730 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.796814 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.797208 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:23Z","lastTransitionTime":"2026-01-26T16:59:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.817378 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v7579" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a77e85f9-b566-4807-bb92-55963c97b93c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba87c9fc35c230bbee201a5176cb467309f0b9aee82dfc81f3b677a15486d02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c03dc794e9c2035f2e1983eacad3e51d76223cb1b82e2f402c73f9453e4bd2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v7579\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:23Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.832845 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af4930ff9cebd62106545fb3da2dfc93f7b591426fe4c85aa2e637b60c935f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:23Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.853990 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c09748c0963bdb50c9552f249d1135ea53e68c52babd788dad050951cb849cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:23Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.876514 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:23Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.889702 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:23Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.899624 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.899660 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.899668 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.899683 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.899692 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:23Z","lastTransitionTime":"2026-01-26T16:59:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.901317 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://627cd6b39fdedd967fffcfc0755439277b2d73016fba1ddc7f5b9d5deba43b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:23Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.915290 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rq622" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a742e7b-c420-46e3-9e96-e9c744af6124\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad7222c9b91c0065a545bc1904d9864a5923cc13bfb6617daeb4a965a830f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8plh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rq622\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:23Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.929451 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ecd87a-c5db-446d-ad3e-cfabbd648c1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89975b8f9428f81ab5d3fb48ced5dd9c837bea2feea3b89f5f7ff8d7d5d15b3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d432159a07ba20bf95f058bf8a597d67cbd8d852519bed035694f8ba3d8ec4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59ba0f131eee14df389391e46bc772894e49e976c3663df5fb3bc98b3cb4a3d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03a01b651d9f66da4dd1f6e0d29ad97c0d6ae46b644c3d997d8dc99476706df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f07438e20bdf71c752bb661084c835341999c561bb5442d75e177223881276f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:58:51Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:58:43.431198 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:58:43.432227 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99450334/tls.crt::/tmp/serving-cert-99450334/tls.key\\\\\\\"\\\\nI0126 16:58:50.828320 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:58:50.830449 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:58:50.830472 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:58:50.830509 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:58:50.830515 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:58:50.834885 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:58:50.834958 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:58:50.834991 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0126 16:58:50.834905 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0126 16:58:50.835015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:58:50.835084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:58:50.835106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:58:50.835116 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0126 16:58:50.837641 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2daa3f13d37b7ae19dfca406b6b5cfdcd4f211287f7d410193f2fee36a24553\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:23Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.942941 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191ad5d41024c88c1a4fbc30f307dd6340fd55b93b00d22f62209d0e82be286f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2a4d8be409a46ffe5702797e56a80023a72a19fb7dc5c49e5f4984cedc600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:23Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.955309 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c823748751e9938f10fab08e33b5fffff5a6d15961ce028204131b7b69a56c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:23Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:23 crc kubenswrapper[4856]: I0126 16:59:23.967745 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:23Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.003596 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.003667 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.003687 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.003712 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.003730 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:24Z","lastTransitionTime":"2026-01-26T16:59:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.015943 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad7b59f9-beb7-49d6-a2d1-e29133e46854\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53a6a6d4222ee5e1ac29fd5957d8cdf3fd42de72c68c85329374ce7afc4004\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v2l7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:24Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.051799 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c166114ee4e41f0a7e4b0590da090e98c319ef6eda0b9611419dfc55ceb139c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c166114ee4e41f0a7e4b0590da090e98c319ef6eda0b9611419dfc55ceb139c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:59:12Z\\\",\\\"message\\\":\\\"etworkPolicy event handler 4 for removal\\\\nI0126 16:59:12.299079 6308 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0126 16:59:12.299145 6308 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0126 16:59:12.299187 6308 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0126 16:59:12.299270 6308 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0126 16:59:12.299298 6308 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0126 16:59:12.299298 6308 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0126 16:59:12.299306 6308 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0126 16:59:12.299340 6308 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0126 16:59:12.299299 6308 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0126 16:59:12.299385 6308 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0126 16:59:12.312081 6308 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0126 16:59:12.312141 6308 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0126 16:59:12.312158 6308 factory.go:656] Stopping watch factory\\\\nI0126 16:59:12.312172 6308 ovnkube.go:599] Stopped ovnkube\\\\nI0126 16:59:12.312201 6308 handler.go:208] Removed *v1.Node event handler 2\\\\nI01\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pxh94_openshift-ovn-kubernetes(ab5b6f50-172b-4535-a0f9-5d103bcab4e7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pxh94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:24Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.066663 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-295wr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12e50462-28e6-4531-ada4-e652310e6cce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-295wr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:24Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.077933 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc864b0d-83bc-4954-9c61-ad650157caff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbde934a6c8acad10ca3ab8206d0ddbd4f7b17e9d304b898a68f4d3b0303bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b48f763d4aff37169399be766d5ab4f7ebbf91f304d139c9022a8556946eb107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0094e662f53c4832a984e05a880021af05ffc4c27f25394c28a070d9ef5490d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb95f623fa4f5f42217649ccce4225f9fe588bb3558da9334eef2b40ddb62486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb95f623fa4f5f42217649ccce4225f9fe588bb3558da9334eef2b40ddb62486\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:24Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.088015 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63c75ede-5170-4db0-811b-5217ef8d72b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da26ddcfc6ccde3c9aabd63bdef3435f9a9eaab8c095bfeef5670c1295576cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54ca3fa13d9e8d442efa93b44a870369f7df3fe7562d77b98528f5c19a751f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xm9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:24Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.105449 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.105497 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.105506 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.105536 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.105549 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:24Z","lastTransitionTime":"2026-01-26T16:59:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.209517 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.209620 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.209661 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.209697 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.209726 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:24Z","lastTransitionTime":"2026-01-26T16:59:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.312869 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.312954 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.312972 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.313002 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.313020 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:24Z","lastTransitionTime":"2026-01-26T16:59:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.395012 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.395087 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.395095 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:59:24 crc kubenswrapper[4856]: E0126 16:59:24.395251 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.395287 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 16:59:24 crc kubenswrapper[4856]: E0126 16:59:24.395441 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:59:24 crc kubenswrapper[4856]: E0126 16:59:24.395617 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 16:59:24 crc kubenswrapper[4856]: E0126 16:59:24.395753 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.415929 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.415997 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.416018 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.416040 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.416057 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:24Z","lastTransitionTime":"2026-01-26T16:59:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.476265 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 13:59:59.608268325 +0000 UTC Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.519209 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.519248 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.519259 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.519320 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.519335 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:24Z","lastTransitionTime":"2026-01-26T16:59:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.622111 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.622187 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.622205 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.622230 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.622249 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:24Z","lastTransitionTime":"2026-01-26T16:59:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.725413 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.725477 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.725496 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.725519 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.725567 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:24Z","lastTransitionTime":"2026-01-26T16:59:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.829567 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.829637 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.829656 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.829685 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.829703 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:24Z","lastTransitionTime":"2026-01-26T16:59:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.933043 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.933094 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.933106 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.933125 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:24 crc kubenswrapper[4856]: I0126 16:59:24.933144 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:24Z","lastTransitionTime":"2026-01-26T16:59:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.036730 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.036774 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.036790 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.036807 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.036819 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:25Z","lastTransitionTime":"2026-01-26T16:59:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.138965 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.139021 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.139033 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.139053 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.139106 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:25Z","lastTransitionTime":"2026-01-26T16:59:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.242222 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.242259 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.242269 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.242286 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.242297 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:25Z","lastTransitionTime":"2026-01-26T16:59:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.345072 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.345139 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.345161 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.345190 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.345213 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:25Z","lastTransitionTime":"2026-01-26T16:59:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.418840 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:25Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.437373 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad7b59f9-beb7-49d6-a2d1-e29133e46854\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53a6a6d4222ee5e1ac29fd5957d8cdf3fd42de72c68c85329374ce7afc4004\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v2l7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:25Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.448679 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.448762 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.448788 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.448821 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.448846 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:25Z","lastTransitionTime":"2026-01-26T16:59:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.464294 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c166114ee4e41f0a7e4b0590da090e98c319ef6eda0b9611419dfc55ceb139c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c166114ee4e41f0a7e4b0590da090e98c319ef6eda0b9611419dfc55ceb139c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:59:12Z\\\",\\\"message\\\":\\\"etworkPolicy event handler 4 for removal\\\\nI0126 16:59:12.299079 6308 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0126 16:59:12.299145 6308 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0126 16:59:12.299187 6308 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0126 16:59:12.299270 6308 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0126 16:59:12.299298 6308 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0126 16:59:12.299298 6308 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0126 16:59:12.299306 6308 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0126 16:59:12.299340 6308 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0126 16:59:12.299299 6308 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0126 16:59:12.299385 6308 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0126 16:59:12.312081 6308 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0126 16:59:12.312141 6308 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0126 16:59:12.312158 6308 factory.go:656] Stopping watch factory\\\\nI0126 16:59:12.312172 6308 ovnkube.go:599] Stopped ovnkube\\\\nI0126 16:59:12.312201 6308 handler.go:208] Removed *v1.Node event handler 2\\\\nI01\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pxh94_openshift-ovn-kubernetes(ab5b6f50-172b-4535-a0f9-5d103bcab4e7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pxh94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:25Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.476656 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 15:22:52.488465242 +0000 UTC Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.481271 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-295wr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12e50462-28e6-4531-ada4-e652310e6cce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-295wr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:25Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.498848 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc864b0d-83bc-4954-9c61-ad650157caff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbde934a6c8acad10ca3ab8206d0ddbd4f7b17e9d304b898a68f4d3b0303bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b48f763d4aff37169399be766d5ab4f7ebbf91f304d139c9022a8556946eb107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0094e662f53c4832a984e05a880021af05ffc4c27f25394c28a070d9ef5490d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb95f623fa4f5f42217649ccce4225f9fe588bb3558da9334eef2b40ddb62486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb95f623fa4f5f42217649ccce4225f9fe588bb3558da9334eef2b40ddb62486\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:25Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.516065 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63c75ede-5170-4db0-811b-5217ef8d72b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da26ddcfc6ccde3c9aabd63bdef3435f9a9eaab8c095bfeef5670c1295576cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54ca3fa13d9e8d442efa93b44a870369f7df3fe7562d77b98528f5c19a751f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xm9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:25Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.562234 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.562323 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.562341 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.562413 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.562447 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:25Z","lastTransitionTime":"2026-01-26T16:59:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.576195 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af4930ff9cebd62106545fb3da2dfc93f7b591426fe4c85aa2e637b60c935f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:25Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.599160 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c09748c0963bdb50c9552f249d1135ea53e68c52babd788dad050951cb849cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:25Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.612637 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:25Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.625252 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v7579" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a77e85f9-b566-4807-bb92-55963c97b93c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba87c9fc35c230bbee201a5176cb467309f0b9aee82dfc81f3b677a15486d02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c03dc794e9c2035f2e1983eacad3e51d76223cb1b82e2f402c73f9453e4bd2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v7579\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:25Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.639652 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ecd87a-c5db-446d-ad3e-cfabbd648c1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89975b8f9428f81ab5d3fb48ced5dd9c837bea2feea3b89f5f7ff8d7d5d15b3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d432159a07ba20bf95f058bf8a597d67cbd8d852519bed035694f8ba3d8ec4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59ba0f131eee14df389391e46bc772894e49e976c3663df5fb3bc98b3cb4a3d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03a01b651d9f66da4dd1f6e0d29ad97c0d6ae46b644c3d997d8dc99476706df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f07438e20bdf71c752bb661084c835341999c561bb5442d75e177223881276f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:58:51Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:58:43.431198 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:58:43.432227 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99450334/tls.crt::/tmp/serving-cert-99450334/tls.key\\\\\\\"\\\\nI0126 16:58:50.828320 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:58:50.830449 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:58:50.830472 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:58:50.830509 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:58:50.830515 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:58:50.834885 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:58:50.834958 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:58:50.834991 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0126 16:58:50.834905 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0126 16:58:50.835015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:58:50.835084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:58:50.835106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:58:50.835116 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0126 16:58:50.837641 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2daa3f13d37b7ae19dfca406b6b5cfdcd4f211287f7d410193f2fee36a24553\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:25Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.652563 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191ad5d41024c88c1a4fbc30f307dd6340fd55b93b00d22f62209d0e82be286f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2a4d8be409a46ffe5702797e56a80023a72a19fb7dc5c49e5f4984cedc600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:25Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.664006 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c823748751e9938f10fab08e33b5fffff5a6d15961ce028204131b7b69a56c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:25Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.665864 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.665901 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.665912 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.665929 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.665941 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:25Z","lastTransitionTime":"2026-01-26T16:59:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.675005 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:25Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.685442 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://627cd6b39fdedd967fffcfc0755439277b2d73016fba1ddc7f5b9d5deba43b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:25Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.700247 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rq622" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a742e7b-c420-46e3-9e96-e9c744af6124\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad7222c9b91c0065a545bc1904d9864a5923cc13bfb6617daeb4a965a830f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8plh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rq622\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:25Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.768676 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.768737 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.768754 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.768783 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.768808 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:25Z","lastTransitionTime":"2026-01-26T16:59:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.871577 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.871640 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.871659 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.871681 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.871698 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:25Z","lastTransitionTime":"2026-01-26T16:59:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.975590 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.975648 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.975666 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.975689 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:25 crc kubenswrapper[4856]: I0126 16:59:25.975703 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:25Z","lastTransitionTime":"2026-01-26T16:59:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.079667 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.079750 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.079775 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.079812 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.079838 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:26Z","lastTransitionTime":"2026-01-26T16:59:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.182956 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.182996 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.183010 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.183028 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.183043 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:26Z","lastTransitionTime":"2026-01-26T16:59:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.202139 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12e50462-28e6-4531-ada4-e652310e6cce-metrics-certs\") pod \"network-metrics-daemon-295wr\" (UID: \"12e50462-28e6-4531-ada4-e652310e6cce\") " pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 16:59:26 crc kubenswrapper[4856]: E0126 16:59:26.202499 4856 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 16:59:26 crc kubenswrapper[4856]: E0126 16:59:26.202666 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12e50462-28e6-4531-ada4-e652310e6cce-metrics-certs podName:12e50462-28e6-4531-ada4-e652310e6cce nodeName:}" failed. No retries permitted until 2026-01-26 16:59:42.202611998 +0000 UTC m=+78.155865979 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/12e50462-28e6-4531-ada4-e652310e6cce-metrics-certs") pod "network-metrics-daemon-295wr" (UID: "12e50462-28e6-4531-ada4-e652310e6cce") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.285213 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.285260 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.285271 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.285288 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.285298 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:26Z","lastTransitionTime":"2026-01-26T16:59:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.387968 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.388352 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.388364 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.388407 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.388425 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:26Z","lastTransitionTime":"2026-01-26T16:59:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.394313 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.394373 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.394313 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 16:59:26 crc kubenswrapper[4856]: E0126 16:59:26.394422 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:59:26 crc kubenswrapper[4856]: E0126 16:59:26.394495 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.394548 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:59:26 crc kubenswrapper[4856]: E0126 16:59:26.394577 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 16:59:26 crc kubenswrapper[4856]: E0126 16:59:26.394641 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.478283 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 15:41:02.743482816 +0000 UTC Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.491774 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.491815 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.491824 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.491839 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.492920 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:26Z","lastTransitionTime":"2026-01-26T16:59:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.596283 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.596337 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.596349 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.596369 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.596382 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:26Z","lastTransitionTime":"2026-01-26T16:59:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.606688 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.606902 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.606949 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.606976 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:59:26 crc kubenswrapper[4856]: E0126 16:59:26.607009 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 16:59:58.606977354 +0000 UTC m=+94.560231335 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.607085 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:59:26 crc kubenswrapper[4856]: E0126 16:59:26.607125 4856 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 16:59:26 crc kubenswrapper[4856]: E0126 16:59:26.607184 4856 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 16:59:26 crc kubenswrapper[4856]: E0126 16:59:26.607222 4856 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 16:59:26 crc kubenswrapper[4856]: E0126 16:59:26.607239 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 16:59:58.60721099 +0000 UTC m=+94.560464981 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 16:59:26 crc kubenswrapper[4856]: E0126 16:59:26.607247 4856 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:59:26 crc kubenswrapper[4856]: E0126 16:59:26.607266 4856 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 16:59:26 crc kubenswrapper[4856]: E0126 16:59:26.607283 4856 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 16:59:26 crc kubenswrapper[4856]: E0126 16:59:26.607294 4856 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:59:26 crc kubenswrapper[4856]: E0126 16:59:26.607298 4856 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 16:59:26 crc kubenswrapper[4856]: E0126 16:59:26.607313 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-26 16:59:58.607299973 +0000 UTC m=+94.560554164 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:59:26 crc kubenswrapper[4856]: E0126 16:59:26.607331 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-26 16:59:58.607324094 +0000 UTC m=+94.560578075 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:59:26 crc kubenswrapper[4856]: E0126 16:59:26.607371 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 16:59:58.607350934 +0000 UTC m=+94.560604905 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.756795 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.756829 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.756838 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.756875 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.756884 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:26Z","lastTransitionTime":"2026-01-26T16:59:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.859487 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.859549 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.859559 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.859578 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.859590 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:26Z","lastTransitionTime":"2026-01-26T16:59:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.962315 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.962364 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.962376 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.962398 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:26 crc kubenswrapper[4856]: I0126 16:59:26.962410 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:26Z","lastTransitionTime":"2026-01-26T16:59:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.064929 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.064970 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.064980 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.064999 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.065011 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:27Z","lastTransitionTime":"2026-01-26T16:59:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.174243 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.174307 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.174319 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.174339 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.174359 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:27Z","lastTransitionTime":"2026-01-26T16:59:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.277208 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.277257 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.277283 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.277309 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.277324 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:27Z","lastTransitionTime":"2026-01-26T16:59:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.378992 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.379063 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.379086 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.379116 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.379140 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:27Z","lastTransitionTime":"2026-01-26T16:59:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.479290 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 19:08:33.920138624 +0000 UTC Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.481169 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.481215 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.481230 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.481251 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.481273 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:27Z","lastTransitionTime":"2026-01-26T16:59:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.583922 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.583986 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.584009 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.584032 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.584047 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:27Z","lastTransitionTime":"2026-01-26T16:59:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.686116 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.686144 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.686154 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.686167 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.686176 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:27Z","lastTransitionTime":"2026-01-26T16:59:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.788192 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.788226 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.788235 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.788249 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.788259 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:27Z","lastTransitionTime":"2026-01-26T16:59:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.890777 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.890825 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.890834 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.890851 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.890861 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:27Z","lastTransitionTime":"2026-01-26T16:59:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.993676 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.993729 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.993747 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.993763 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:27 crc kubenswrapper[4856]: I0126 16:59:27.993773 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:27Z","lastTransitionTime":"2026-01-26T16:59:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.097653 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.097698 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.097709 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.097725 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.097736 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:28Z","lastTransitionTime":"2026-01-26T16:59:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.200676 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.200713 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.200722 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.200736 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.200746 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:28Z","lastTransitionTime":"2026-01-26T16:59:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.303629 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.303690 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.303707 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.303735 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.303753 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:28Z","lastTransitionTime":"2026-01-26T16:59:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.395004 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.395038 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.395106 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.395062 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:59:28 crc kubenswrapper[4856]: E0126 16:59:28.395190 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 16:59:28 crc kubenswrapper[4856]: E0126 16:59:28.395264 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:59:28 crc kubenswrapper[4856]: E0126 16:59:28.395370 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:59:28 crc kubenswrapper[4856]: E0126 16:59:28.395804 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.396437 4856 scope.go:117] "RemoveContainer" containerID="7c166114ee4e41f0a7e4b0590da090e98c319ef6eda0b9611419dfc55ceb139c" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.406676 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.406712 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.406723 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.406740 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.406751 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:28Z","lastTransitionTime":"2026-01-26T16:59:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.479541 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 20:08:43.544369216 +0000 UTC Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.508626 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.508674 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.508683 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.508702 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.508715 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:28Z","lastTransitionTime":"2026-01-26T16:59:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.611152 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.611192 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.611205 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.611224 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.611241 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:28Z","lastTransitionTime":"2026-01-26T16:59:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.714379 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.714420 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.714431 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.714448 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.714463 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:28Z","lastTransitionTime":"2026-01-26T16:59:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.816436 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.816499 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.816513 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.816556 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.816575 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:28Z","lastTransitionTime":"2026-01-26T16:59:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.858041 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.858088 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.858101 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.858118 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.858130 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:28Z","lastTransitionTime":"2026-01-26T16:59:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:28 crc kubenswrapper[4856]: E0126 16:59:28.871657 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17523591-a778-4a97-aeab-8a7a93101850\\\",\\\"systemUUID\\\":\\\"ca45d056-99cb-4442-8a44-7e899628ecb2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:28Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.875697 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.875752 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.875768 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.875787 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.875798 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:28Z","lastTransitionTime":"2026-01-26T16:59:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:28 crc kubenswrapper[4856]: E0126 16:59:28.887860 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17523591-a778-4a97-aeab-8a7a93101850\\\",\\\"systemUUID\\\":\\\"ca45d056-99cb-4442-8a44-7e899628ecb2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:28Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.891093 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.891118 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.891140 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.891154 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.891164 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:28Z","lastTransitionTime":"2026-01-26T16:59:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:28 crc kubenswrapper[4856]: E0126 16:59:28.908003 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17523591-a778-4a97-aeab-8a7a93101850\\\",\\\"systemUUID\\\":\\\"ca45d056-99cb-4442-8a44-7e899628ecb2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:28Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.911604 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.911638 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.911649 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.911667 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.911679 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:28Z","lastTransitionTime":"2026-01-26T16:59:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:28 crc kubenswrapper[4856]: E0126 16:59:28.927408 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17523591-a778-4a97-aeab-8a7a93101850\\\",\\\"systemUUID\\\":\\\"ca45d056-99cb-4442-8a44-7e899628ecb2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:28Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.930714 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.930735 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.930743 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.930756 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:28 crc kubenswrapper[4856]: I0126 16:59:28.930765 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:28Z","lastTransitionTime":"2026-01-26T16:59:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:28 crc kubenswrapper[4856]: E0126 16:59:28.943258 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17523591-a778-4a97-aeab-8a7a93101850\\\",\\\"systemUUID\\\":\\\"ca45d056-99cb-4442-8a44-7e899628ecb2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:28Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:28 crc kubenswrapper[4856]: E0126 16:59:28.943449 4856 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.039401 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.039458 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.039469 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.039485 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.039496 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:29Z","lastTransitionTime":"2026-01-26T16:59:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.141798 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.141840 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.141851 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.141867 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.141881 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:29Z","lastTransitionTime":"2026-01-26T16:59:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.244182 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.244221 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.244232 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.244250 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.244261 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:29Z","lastTransitionTime":"2026-01-26T16:59:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.304812 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pxh94_ab5b6f50-172b-4535-a0f9-5d103bcab4e7/ovnkube-controller/1.log" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.307192 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" event={"ID":"ab5b6f50-172b-4535-a0f9-5d103bcab4e7","Type":"ContainerStarted","Data":"71cb66e6f52823e7aa63f88ba1d153fde73816120aa75d4a6b910937303d2b9e"} Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.307642 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.346431 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.346682 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.346791 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.346912 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.347022 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:29Z","lastTransitionTime":"2026-01-26T16:59:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.356155 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-295wr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12e50462-28e6-4531-ada4-e652310e6cce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-295wr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:29Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.373361 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:29Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.389690 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad7b59f9-beb7-49d6-a2d1-e29133e46854\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53a6a6d4222ee5e1ac29fd5957d8cdf3fd42de72c68c85329374ce7afc4004\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v2l7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:29Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.416748 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71cb66e6f52823e7aa63f88ba1d153fde73816120aa75d4a6b910937303d2b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c166114ee4e41f0a7e4b0590da090e98c319ef6eda0b9611419dfc55ceb139c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:59:12Z\\\",\\\"message\\\":\\\"etworkPolicy event handler 4 for removal\\\\nI0126 16:59:12.299079 6308 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0126 16:59:12.299145 6308 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0126 16:59:12.299187 6308 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0126 16:59:12.299270 6308 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0126 16:59:12.299298 6308 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0126 16:59:12.299298 6308 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0126 16:59:12.299306 6308 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0126 16:59:12.299340 6308 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0126 16:59:12.299299 6308 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0126 16:59:12.299385 6308 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0126 16:59:12.312081 6308 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0126 16:59:12.312141 6308 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0126 16:59:12.312158 6308 factory.go:656] Stopping watch factory\\\\nI0126 16:59:12.312172 6308 ovnkube.go:599] Stopped ovnkube\\\\nI0126 16:59:12.312201 6308 handler.go:208] Removed *v1.Node event handler 2\\\\nI01\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pxh94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:29Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.430870 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc864b0d-83bc-4954-9c61-ad650157caff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbde934a6c8acad10ca3ab8206d0ddbd4f7b17e9d304b898a68f4d3b0303bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b48f763d4aff37169399be766d5ab4f7ebbf91f304d139c9022a8556946eb107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0094e662f53c4832a984e05a880021af05ffc4c27f25394c28a070d9ef5490d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb95f623fa4f5f42217649ccce4225f9fe588bb3558da9334eef2b40ddb62486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb95f623fa4f5f42217649ccce4225f9fe588bb3558da9334eef2b40ddb62486\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:29Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.444814 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63c75ede-5170-4db0-811b-5217ef8d72b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da26ddcfc6ccde3c9aabd63bdef3435f9a9eaab8c095bfeef5670c1295576cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54ca3fa13d9e8d442efa93b44a870369f7df3fe7562d77b98528f5c19a751f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xm9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:29Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.452166 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.452202 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.452213 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.452230 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.452240 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:29Z","lastTransitionTime":"2026-01-26T16:59:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.460499 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:29Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.488927 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v7579" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a77e85f9-b566-4807-bb92-55963c97b93c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba87c9fc35c230bbee201a5176cb467309f0b9aee82dfc81f3b677a15486d02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c03dc794e9c2035f2e1983eacad3e51d76223cb1b82e2f402c73f9453e4bd2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v7579\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:29Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.504412 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af4930ff9cebd62106545fb3da2dfc93f7b591426fe4c85aa2e637b60c935f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:29Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.519454 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c09748c0963bdb50c9552f249d1135ea53e68c52babd788dad050951cb849cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:29Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.527652 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-08 02:57:30.885241647 +0000 UTC Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.537433 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c823748751e9938f10fab08e33b5fffff5a6d15961ce028204131b7b69a56c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:29Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.555235 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.555275 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.555288 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.555305 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.555315 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:29Z","lastTransitionTime":"2026-01-26T16:59:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.558147 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:29Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.593336 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://627cd6b39fdedd967fffcfc0755439277b2d73016fba1ddc7f5b9d5deba43b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:29Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.614746 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rq622" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a742e7b-c420-46e3-9e96-e9c744af6124\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad7222c9b91c0065a545bc1904d9864a5923cc13bfb6617daeb4a965a830f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8plh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rq622\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:29Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.639028 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ecd87a-c5db-446d-ad3e-cfabbd648c1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89975b8f9428f81ab5d3fb48ced5dd9c837bea2feea3b89f5f7ff8d7d5d15b3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d432159a07ba20bf95f058bf8a597d67cbd8d852519bed035694f8ba3d8ec4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59ba0f131eee14df389391e46bc772894e49e976c3663df5fb3bc98b3cb4a3d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03a01b651d9f66da4dd1f6e0d29ad97c0d6ae46b644c3d997d8dc99476706df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f07438e20bdf71c752bb661084c835341999c561bb5442d75e177223881276f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:58:51Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:58:43.431198 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:58:43.432227 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99450334/tls.crt::/tmp/serving-cert-99450334/tls.key\\\\\\\"\\\\nI0126 16:58:50.828320 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:58:50.830449 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:58:50.830472 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:58:50.830509 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:58:50.830515 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:58:50.834885 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:58:50.834958 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:58:50.834991 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0126 16:58:50.834905 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0126 16:58:50.835015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:58:50.835084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:58:50.835106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:58:50.835116 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0126 16:58:50.837641 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2daa3f13d37b7ae19dfca406b6b5cfdcd4f211287f7d410193f2fee36a24553\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:29Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.655952 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191ad5d41024c88c1a4fbc30f307dd6340fd55b93b00d22f62209d0e82be286f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2a4d8be409a46ffe5702797e56a80023a72a19fb7dc5c49e5f4984cedc600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:29Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.658577 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.658629 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.658644 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.658668 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.658683 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:29Z","lastTransitionTime":"2026-01-26T16:59:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.761708 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.761767 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.761778 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.761803 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.761817 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:29Z","lastTransitionTime":"2026-01-26T16:59:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.864577 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.864624 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.864634 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.864651 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.864662 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:29Z","lastTransitionTime":"2026-01-26T16:59:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.967312 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.967658 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.967749 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.967835 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:29 crc kubenswrapper[4856]: I0126 16:59:29.967903 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:29Z","lastTransitionTime":"2026-01-26T16:59:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.071690 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.071742 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.071755 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.071773 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.071786 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:30Z","lastTransitionTime":"2026-01-26T16:59:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.175121 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.175164 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.175175 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.175190 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.175201 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:30Z","lastTransitionTime":"2026-01-26T16:59:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.511966 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:59:30 crc kubenswrapper[4856]: E0126 16:59:30.512098 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.512169 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.512244 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.512271 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:59:30 crc kubenswrapper[4856]: E0126 16:59:30.512328 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 16:59:30 crc kubenswrapper[4856]: E0126 16:59:30.512777 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:59:30 crc kubenswrapper[4856]: E0126 16:59:30.512823 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.516927 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.516954 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.516963 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.516978 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.516988 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:30Z","lastTransitionTime":"2026-01-26T16:59:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.519202 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pxh94_ab5b6f50-172b-4535-a0f9-5d103bcab4e7/ovnkube-controller/2.log" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.520269 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pxh94_ab5b6f50-172b-4535-a0f9-5d103bcab4e7/ovnkube-controller/1.log" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.522220 4856 generic.go:334] "Generic (PLEG): container finished" podID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerID="71cb66e6f52823e7aa63f88ba1d153fde73816120aa75d4a6b910937303d2b9e" exitCode=1 Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.522254 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" event={"ID":"ab5b6f50-172b-4535-a0f9-5d103bcab4e7","Type":"ContainerDied","Data":"71cb66e6f52823e7aa63f88ba1d153fde73816120aa75d4a6b910937303d2b9e"} Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.522288 4856 scope.go:117] "RemoveContainer" containerID="7c166114ee4e41f0a7e4b0590da090e98c319ef6eda0b9611419dfc55ceb139c" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.522867 4856 scope.go:117] "RemoveContainer" containerID="71cb66e6f52823e7aa63f88ba1d153fde73816120aa75d4a6b910937303d2b9e" Jan 26 16:59:30 crc kubenswrapper[4856]: E0126 16:59:30.522999 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pxh94_openshift-ovn-kubernetes(ab5b6f50-172b-4535-a0f9-5d103bcab4e7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.528159 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 06:33:25.34641691 +0000 UTC Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.540327 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rq622" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a742e7b-c420-46e3-9e96-e9c744af6124\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad7222c9b91c0065a545bc1904d9864a5923cc13bfb6617daeb4a965a830f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8plh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rq622\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:30Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.557590 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ecd87a-c5db-446d-ad3e-cfabbd648c1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89975b8f9428f81ab5d3fb48ced5dd9c837bea2feea3b89f5f7ff8d7d5d15b3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d432159a07ba20bf95f058bf8a597d67cbd8d852519bed035694f8ba3d8ec4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59ba0f131eee14df389391e46bc772894e49e976c3663df5fb3bc98b3cb4a3d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03a01b651d9f66da4dd1f6e0d29ad97c0d6ae46b644c3d997d8dc99476706df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f07438e20bdf71c752bb661084c835341999c561bb5442d75e177223881276f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:58:51Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:58:43.431198 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:58:43.432227 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99450334/tls.crt::/tmp/serving-cert-99450334/tls.key\\\\\\\"\\\\nI0126 16:58:50.828320 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:58:50.830449 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:58:50.830472 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:58:50.830509 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:58:50.830515 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:58:50.834885 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:58:50.834958 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:58:50.834991 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0126 16:58:50.834905 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0126 16:58:50.835015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:58:50.835084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:58:50.835106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:58:50.835116 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0126 16:58:50.837641 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2daa3f13d37b7ae19dfca406b6b5cfdcd4f211287f7d410193f2fee36a24553\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:30Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.572871 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191ad5d41024c88c1a4fbc30f307dd6340fd55b93b00d22f62209d0e82be286f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2a4d8be409a46ffe5702797e56a80023a72a19fb7dc5c49e5f4984cedc600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:30Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.586949 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c823748751e9938f10fab08e33b5fffff5a6d15961ce028204131b7b69a56c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:30Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.602838 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:30Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.614059 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://627cd6b39fdedd967fffcfc0755439277b2d73016fba1ddc7f5b9d5deba43b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:30Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.619977 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.620015 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.620026 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.620041 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.620056 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:30Z","lastTransitionTime":"2026-01-26T16:59:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.631238 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:30Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.650402 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad7b59f9-beb7-49d6-a2d1-e29133e46854\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53a6a6d4222ee5e1ac29fd5957d8cdf3fd42de72c68c85329374ce7afc4004\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v2l7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:30Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.674399 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71cb66e6f52823e7aa63f88ba1d153fde73816120aa75d4a6b910937303d2b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c166114ee4e41f0a7e4b0590da090e98c319ef6eda0b9611419dfc55ceb139c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:59:12Z\\\",\\\"message\\\":\\\"etworkPolicy event handler 4 for removal\\\\nI0126 16:59:12.299079 6308 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0126 16:59:12.299145 6308 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0126 16:59:12.299187 6308 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0126 16:59:12.299270 6308 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0126 16:59:12.299298 6308 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0126 16:59:12.299298 6308 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0126 16:59:12.299306 6308 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0126 16:59:12.299340 6308 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0126 16:59:12.299299 6308 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0126 16:59:12.299385 6308 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0126 16:59:12.312081 6308 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0126 16:59:12.312141 6308 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0126 16:59:12.312158 6308 factory.go:656] Stopping watch factory\\\\nI0126 16:59:12.312172 6308 ovnkube.go:599] Stopped ovnkube\\\\nI0126 16:59:12.312201 6308 handler.go:208] Removed *v1.Node event handler 2\\\\nI01\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71cb66e6f52823e7aa63f88ba1d153fde73816120aa75d4a6b910937303d2b9e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:59:29Z\\\",\\\"message\\\":\\\"n-kubernetes/ovnkube-node-pxh94\\\\nI0126 16:59:29.834990 6550 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-t4fq2\\\\nI0126 16:59:29.834996 6550 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-t4fq2\\\\nI0126 16:59:29.835005 6550 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-t4fq2 in node crc\\\\nI0126 16:59:29.835026 6550 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0126 16:59:29.835056 6550 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0126 16:59:29.835079 6550 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0126 16:59:29.835090 6550 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nF0126 16:59:29.835097 6550 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pxh94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:30Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.689293 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-295wr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12e50462-28e6-4531-ada4-e652310e6cce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-295wr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:30Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.704582 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc864b0d-83bc-4954-9c61-ad650157caff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbde934a6c8acad10ca3ab8206d0ddbd4f7b17e9d304b898a68f4d3b0303bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b48f763d4aff37169399be766d5ab4f7ebbf91f304d139c9022a8556946eb107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0094e662f53c4832a984e05a880021af05ffc4c27f25394c28a070d9ef5490d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb95f623fa4f5f42217649ccce4225f9fe588bb3558da9334eef2b40ddb62486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb95f623fa4f5f42217649ccce4225f9fe588bb3558da9334eef2b40ddb62486\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:30Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.718734 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63c75ede-5170-4db0-811b-5217ef8d72b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da26ddcfc6ccde3c9aabd63bdef3435f9a9eaab8c095bfeef5670c1295576cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54ca3fa13d9e8d442efa93b44a870369f7df3fe7562d77b98528f5c19a751f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xm9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:30Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.722560 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.722598 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.722608 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.722624 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.722634 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:30Z","lastTransitionTime":"2026-01-26T16:59:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.732746 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af4930ff9cebd62106545fb3da2dfc93f7b591426fe4c85aa2e637b60c935f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:30Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.750450 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c09748c0963bdb50c9552f249d1135ea53e68c52babd788dad050951cb849cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:30Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.764769 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:30Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.777515 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v7579" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a77e85f9-b566-4807-bb92-55963c97b93c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba87c9fc35c230bbee201a5176cb467309f0b9aee82dfc81f3b677a15486d02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c03dc794e9c2035f2e1983eacad3e51d76223cb1b82e2f402c73f9453e4bd2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v7579\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:30Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.824464 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.824488 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.824495 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.824510 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.824518 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:30Z","lastTransitionTime":"2026-01-26T16:59:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.927394 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.927424 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.927433 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.927446 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:30 crc kubenswrapper[4856]: I0126 16:59:30.927455 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:30Z","lastTransitionTime":"2026-01-26T16:59:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.029635 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.029694 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.029710 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.029731 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.029745 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:31Z","lastTransitionTime":"2026-01-26T16:59:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.132597 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.132645 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.132655 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.132671 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.132682 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:31Z","lastTransitionTime":"2026-01-26T16:59:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.235318 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.235473 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.235516 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.235575 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.235591 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:31Z","lastTransitionTime":"2026-01-26T16:59:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.338218 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.338252 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.338259 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.338272 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.338281 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:31Z","lastTransitionTime":"2026-01-26T16:59:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.440950 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.440999 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.441011 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.441029 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.441040 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:31Z","lastTransitionTime":"2026-01-26T16:59:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.526708 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pxh94_ab5b6f50-172b-4535-a0f9-5d103bcab4e7/ovnkube-controller/2.log" Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.528261 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 09:57:18.474251678 +0000 UTC Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.530106 4856 scope.go:117] "RemoveContainer" containerID="71cb66e6f52823e7aa63f88ba1d153fde73816120aa75d4a6b910937303d2b9e" Jan 26 16:59:31 crc kubenswrapper[4856]: E0126 16:59:31.530328 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pxh94_openshift-ovn-kubernetes(ab5b6f50-172b-4535-a0f9-5d103bcab4e7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.542753 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.542793 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.542805 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.542826 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.542837 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:31Z","lastTransitionTime":"2026-01-26T16:59:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.545766 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:31Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.561827 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad7b59f9-beb7-49d6-a2d1-e29133e46854\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53a6a6d4222ee5e1ac29fd5957d8cdf3fd42de72c68c85329374ce7afc4004\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v2l7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:31Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.924702 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.924732 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.924742 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.924755 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.924764 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:31Z","lastTransitionTime":"2026-01-26T16:59:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.955745 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71cb66e6f52823e7aa63f88ba1d153fde73816120aa75d4a6b910937303d2b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71cb66e6f52823e7aa63f88ba1d153fde73816120aa75d4a6b910937303d2b9e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:59:29Z\\\",\\\"message\\\":\\\"n-kubernetes/ovnkube-node-pxh94\\\\nI0126 16:59:29.834990 6550 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-t4fq2\\\\nI0126 16:59:29.834996 6550 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-t4fq2\\\\nI0126 16:59:29.835005 6550 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-t4fq2 in node crc\\\\nI0126 16:59:29.835026 6550 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0126 16:59:29.835056 6550 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0126 16:59:29.835079 6550 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0126 16:59:29.835090 6550 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nF0126 16:59:29.835097 6550 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pxh94_openshift-ovn-kubernetes(ab5b6f50-172b-4535-a0f9-5d103bcab4e7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pxh94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:31Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.968813 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-295wr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12e50462-28e6-4531-ada4-e652310e6cce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-295wr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:31Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.984725 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc864b0d-83bc-4954-9c61-ad650157caff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbde934a6c8acad10ca3ab8206d0ddbd4f7b17e9d304b898a68f4d3b0303bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b48f763d4aff37169399be766d5ab4f7ebbf91f304d139c9022a8556946eb107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0094e662f53c4832a984e05a880021af05ffc4c27f25394c28a070d9ef5490d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb95f623fa4f5f42217649ccce4225f9fe588bb3558da9334eef2b40ddb62486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb95f623fa4f5f42217649ccce4225f9fe588bb3558da9334eef2b40ddb62486\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:31Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:31 crc kubenswrapper[4856]: I0126 16:59:31.995794 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63c75ede-5170-4db0-811b-5217ef8d72b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da26ddcfc6ccde3c9aabd63bdef3435f9a9eaab8c095bfeef5670c1295576cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54ca3fa13d9e8d442efa93b44a870369f7df3fe7562d77b98528f5c19a751f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xm9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:31Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.006174 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af4930ff9cebd62106545fb3da2dfc93f7b591426fe4c85aa2e637b60c935f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:32Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.018911 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c09748c0963bdb50c9552f249d1135ea53e68c52babd788dad050951cb849cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:32Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.026645 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.026680 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.026691 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.026709 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.026722 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:32Z","lastTransitionTime":"2026-01-26T16:59:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.033665 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:32Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.045201 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v7579" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a77e85f9-b566-4807-bb92-55963c97b93c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba87c9fc35c230bbee201a5176cb467309f0b9aee82dfc81f3b677a15486d02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c03dc794e9c2035f2e1983eacad3e51d76223cb1b82e2f402c73f9453e4bd2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v7579\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:32Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.060010 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ecd87a-c5db-446d-ad3e-cfabbd648c1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89975b8f9428f81ab5d3fb48ced5dd9c837bea2feea3b89f5f7ff8d7d5d15b3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d432159a07ba20bf95f058bf8a597d67cbd8d852519bed035694f8ba3d8ec4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59ba0f131eee14df389391e46bc772894e49e976c3663df5fb3bc98b3cb4a3d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03a01b651d9f66da4dd1f6e0d29ad97c0d6ae46b644c3d997d8dc99476706df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f07438e20bdf71c752bb661084c835341999c561bb5442d75e177223881276f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:58:51Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:58:43.431198 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:58:43.432227 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99450334/tls.crt::/tmp/serving-cert-99450334/tls.key\\\\\\\"\\\\nI0126 16:58:50.828320 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:58:50.830449 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:58:50.830472 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:58:50.830509 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:58:50.830515 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:58:50.834885 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:58:50.834958 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:58:50.834991 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0126 16:58:50.834905 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0126 16:58:50.835015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:58:50.835084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:58:50.835106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:58:50.835116 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0126 16:58:50.837641 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2daa3f13d37b7ae19dfca406b6b5cfdcd4f211287f7d410193f2fee36a24553\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:32Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.075013 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191ad5d41024c88c1a4fbc30f307dd6340fd55b93b00d22f62209d0e82be286f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2a4d8be409a46ffe5702797e56a80023a72a19fb7dc5c49e5f4984cedc600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:32Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.089616 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c823748751e9938f10fab08e33b5fffff5a6d15961ce028204131b7b69a56c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:32Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.104920 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:32Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.118714 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://627cd6b39fdedd967fffcfc0755439277b2d73016fba1ddc7f5b9d5deba43b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:32Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.128924 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.128960 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.128969 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.128987 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.128998 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:32Z","lastTransitionTime":"2026-01-26T16:59:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.136177 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rq622" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a742e7b-c420-46e3-9e96-e9c744af6124\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad7222c9b91c0065a545bc1904d9864a5923cc13bfb6617daeb4a965a830f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8plh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rq622\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:32Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.231488 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.231517 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.231542 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.231554 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.231564 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:32Z","lastTransitionTime":"2026-01-26T16:59:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.333732 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.333780 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.333791 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.333808 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.333819 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:32Z","lastTransitionTime":"2026-01-26T16:59:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.395062 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.395066 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.395117 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.395192 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:59:32 crc kubenswrapper[4856]: E0126 16:59:32.395339 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 16:59:32 crc kubenswrapper[4856]: E0126 16:59:32.395452 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:59:32 crc kubenswrapper[4856]: E0126 16:59:32.395590 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:59:32 crc kubenswrapper[4856]: E0126 16:59:32.395694 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.436422 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.436469 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.436484 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.436502 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.436517 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:32Z","lastTransitionTime":"2026-01-26T16:59:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.528383 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 14:12:46.832808478 +0000 UTC Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.538786 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.538814 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.538821 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.538834 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.538843 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:32Z","lastTransitionTime":"2026-01-26T16:59:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.645656 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.645716 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.645730 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.645755 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.645772 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:32Z","lastTransitionTime":"2026-01-26T16:59:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.748470 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.748503 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.748520 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.748559 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.748569 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:32Z","lastTransitionTime":"2026-01-26T16:59:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.851488 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.851551 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.851564 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.851581 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.851593 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:32Z","lastTransitionTime":"2026-01-26T16:59:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.954935 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.954969 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.954988 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.955005 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:32 crc kubenswrapper[4856]: I0126 16:59:32.955015 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:32Z","lastTransitionTime":"2026-01-26T16:59:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.057299 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.057328 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.057336 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.057361 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.057370 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:33Z","lastTransitionTime":"2026-01-26T16:59:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.159660 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.159702 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.159719 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.159736 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.159745 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:33Z","lastTransitionTime":"2026-01-26T16:59:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.262844 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.262872 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.262879 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.262895 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.262906 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:33Z","lastTransitionTime":"2026-01-26T16:59:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.365511 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.365595 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.365607 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.365624 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.365636 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:33Z","lastTransitionTime":"2026-01-26T16:59:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.468494 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.468557 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.468569 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.468587 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.468597 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:33Z","lastTransitionTime":"2026-01-26T16:59:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.529167 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 17:05:39.252262224 +0000 UTC Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.571610 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.571651 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.571662 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.571679 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.571695 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:33Z","lastTransitionTime":"2026-01-26T16:59:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.674812 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.674860 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.674871 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.674885 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.674894 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:33Z","lastTransitionTime":"2026-01-26T16:59:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.777012 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.777052 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.777069 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.777085 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.777095 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:33Z","lastTransitionTime":"2026-01-26T16:59:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.879101 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.879166 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.879179 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.879195 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.879209 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:33Z","lastTransitionTime":"2026-01-26T16:59:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.982003 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.982075 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.982088 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.982104 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:33 crc kubenswrapper[4856]: I0126 16:59:33.982116 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:33Z","lastTransitionTime":"2026-01-26T16:59:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.086031 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.086127 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.086155 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.086221 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.086247 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:34Z","lastTransitionTime":"2026-01-26T16:59:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.189226 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.189276 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.189289 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.189308 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.189319 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:34Z","lastTransitionTime":"2026-01-26T16:59:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.291901 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.291933 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.291942 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.291955 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.291964 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:34Z","lastTransitionTime":"2026-01-26T16:59:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.394169 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.394213 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.394240 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.394174 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:59:34 crc kubenswrapper[4856]: E0126 16:59:34.394285 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:59:34 crc kubenswrapper[4856]: E0126 16:59:34.394365 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.394439 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.394459 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.394468 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:34 crc kubenswrapper[4856]: E0126 16:59:34.394471 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.394485 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.394504 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:34Z","lastTransitionTime":"2026-01-26T16:59:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:34 crc kubenswrapper[4856]: E0126 16:59:34.394560 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.497661 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.497695 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.497707 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.497723 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.497736 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:34Z","lastTransitionTime":"2026-01-26T16:59:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.530243 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 06:22:24.079556356 +0000 UTC Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.600100 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.600189 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.600201 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.600239 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.600251 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:34Z","lastTransitionTime":"2026-01-26T16:59:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.702661 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.702714 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.702729 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.702749 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.702765 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:34Z","lastTransitionTime":"2026-01-26T16:59:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.804578 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.804621 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.804641 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.804659 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.804672 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:34Z","lastTransitionTime":"2026-01-26T16:59:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.907641 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.907676 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.907688 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.907706 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:34 crc kubenswrapper[4856]: I0126 16:59:34.907715 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:34Z","lastTransitionTime":"2026-01-26T16:59:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.010200 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.010272 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.010286 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.010326 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.010340 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:35Z","lastTransitionTime":"2026-01-26T16:59:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.113503 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.113578 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.113587 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.113604 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.113614 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:35Z","lastTransitionTime":"2026-01-26T16:59:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.216463 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.216589 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.216606 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.216665 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.216679 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:35Z","lastTransitionTime":"2026-01-26T16:59:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.319852 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.319904 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.319914 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.319929 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.319940 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:35Z","lastTransitionTime":"2026-01-26T16:59:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.413194 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:35Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.422379 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.422456 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.422469 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.422491 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.422504 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:35Z","lastTransitionTime":"2026-01-26T16:59:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.435977 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad7b59f9-beb7-49d6-a2d1-e29133e46854\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53a6a6d4222ee5e1ac29fd5957d8cdf3fd42de72c68c85329374ce7afc4004\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v2l7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:35Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.461125 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71cb66e6f52823e7aa63f88ba1d153fde73816120aa75d4a6b910937303d2b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71cb66e6f52823e7aa63f88ba1d153fde73816120aa75d4a6b910937303d2b9e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:59:29Z\\\",\\\"message\\\":\\\"n-kubernetes/ovnkube-node-pxh94\\\\nI0126 16:59:29.834990 6550 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-t4fq2\\\\nI0126 16:59:29.834996 6550 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-t4fq2\\\\nI0126 16:59:29.835005 6550 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-t4fq2 in node crc\\\\nI0126 16:59:29.835026 6550 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0126 16:59:29.835056 6550 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0126 16:59:29.835079 6550 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0126 16:59:29.835090 6550 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nF0126 16:59:29.835097 6550 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pxh94_openshift-ovn-kubernetes(ab5b6f50-172b-4535-a0f9-5d103bcab4e7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pxh94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:35Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.479884 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-295wr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12e50462-28e6-4531-ada4-e652310e6cce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-295wr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:35Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.496953 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc864b0d-83bc-4954-9c61-ad650157caff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbde934a6c8acad10ca3ab8206d0ddbd4f7b17e9d304b898a68f4d3b0303bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b48f763d4aff37169399be766d5ab4f7ebbf91f304d139c9022a8556946eb107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0094e662f53c4832a984e05a880021af05ffc4c27f25394c28a070d9ef5490d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb95f623fa4f5f42217649ccce4225f9fe588bb3558da9334eef2b40ddb62486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb95f623fa4f5f42217649ccce4225f9fe588bb3558da9334eef2b40ddb62486\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:35Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.510411 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63c75ede-5170-4db0-811b-5217ef8d72b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da26ddcfc6ccde3c9aabd63bdef3435f9a9eaab8c095bfeef5670c1295576cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54ca3fa13d9e8d442efa93b44a870369f7df3fe7562d77b98528f5c19a751f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xm9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:35Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.520280 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af4930ff9cebd62106545fb3da2dfc93f7b591426fe4c85aa2e637b60c935f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:35Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.524823 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.524855 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.524865 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.524912 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.524923 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:35Z","lastTransitionTime":"2026-01-26T16:59:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.531112 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 16:59:59.499715133 +0000 UTC Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.721874 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.721939 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.721952 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.721968 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.722024 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:35Z","lastTransitionTime":"2026-01-26T16:59:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.724219 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c09748c0963bdb50c9552f249d1135ea53e68c52babd788dad050951cb849cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:35Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.744298 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:35Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.759807 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v7579" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a77e85f9-b566-4807-bb92-55963c97b93c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba87c9fc35c230bbee201a5176cb467309f0b9aee82dfc81f3b677a15486d02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c03dc794e9c2035f2e1983eacad3e51d76223cb1b82e2f402c73f9453e4bd2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v7579\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:35Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.774165 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://627cd6b39fdedd967fffcfc0755439277b2d73016fba1ddc7f5b9d5deba43b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:35Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.792520 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rq622" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a742e7b-c420-46e3-9e96-e9c744af6124\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad7222c9b91c0065a545bc1904d9864a5923cc13bfb6617daeb4a965a830f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8plh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rq622\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:35Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.812460 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ecd87a-c5db-446d-ad3e-cfabbd648c1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89975b8f9428f81ab5d3fb48ced5dd9c837bea2feea3b89f5f7ff8d7d5d15b3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d432159a07ba20bf95f058bf8a597d67cbd8d852519bed035694f8ba3d8ec4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59ba0f131eee14df389391e46bc772894e49e976c3663df5fb3bc98b3cb4a3d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03a01b651d9f66da4dd1f6e0d29ad97c0d6ae46b644c3d997d8dc99476706df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f07438e20bdf71c752bb661084c835341999c561bb5442d75e177223881276f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:58:51Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:58:43.431198 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:58:43.432227 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99450334/tls.crt::/tmp/serving-cert-99450334/tls.key\\\\\\\"\\\\nI0126 16:58:50.828320 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:58:50.830449 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:58:50.830472 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:58:50.830509 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:58:50.830515 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:58:50.834885 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:58:50.834958 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:58:50.834991 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0126 16:58:50.834905 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0126 16:58:50.835015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:58:50.835084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:58:50.835106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:58:50.835116 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0126 16:58:50.837641 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2daa3f13d37b7ae19dfca406b6b5cfdcd4f211287f7d410193f2fee36a24553\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:35Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.824023 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.824045 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.824054 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.824068 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.824079 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:35Z","lastTransitionTime":"2026-01-26T16:59:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.829836 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191ad5d41024c88c1a4fbc30f307dd6340fd55b93b00d22f62209d0e82be286f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2a4d8be409a46ffe5702797e56a80023a72a19fb7dc5c49e5f4984cedc600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:35Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.844280 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c823748751e9938f10fab08e33b5fffff5a6d15961ce028204131b7b69a56c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:35Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.858460 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:35Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.926627 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.926670 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.926679 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.926696 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:35 crc kubenswrapper[4856]: I0126 16:59:35.926706 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:35Z","lastTransitionTime":"2026-01-26T16:59:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.029983 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.030031 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.030041 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.030059 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.030070 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:36Z","lastTransitionTime":"2026-01-26T16:59:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.133142 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.133181 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.133189 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.133204 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.133213 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:36Z","lastTransitionTime":"2026-01-26T16:59:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.235491 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.235568 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.235584 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.235606 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.235620 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:36Z","lastTransitionTime":"2026-01-26T16:59:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.338059 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.338106 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.338113 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.338127 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.338136 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:36Z","lastTransitionTime":"2026-01-26T16:59:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.394557 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.394597 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.394564 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:59:36 crc kubenswrapper[4856]: E0126 16:59:36.394715 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.394564 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 16:59:36 crc kubenswrapper[4856]: E0126 16:59:36.394830 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:59:36 crc kubenswrapper[4856]: E0126 16:59:36.394927 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 16:59:36 crc kubenswrapper[4856]: E0126 16:59:36.394985 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.439867 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.439911 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.439922 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.439938 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.439949 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:36Z","lastTransitionTime":"2026-01-26T16:59:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.531768 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 09:43:06.140636051 +0000 UTC Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.542545 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.542572 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.542582 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.542595 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.542605 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:36Z","lastTransitionTime":"2026-01-26T16:59:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.645348 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.645391 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.645400 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.645415 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.645426 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:36Z","lastTransitionTime":"2026-01-26T16:59:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.747563 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.747605 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.747617 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.747635 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.747650 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:36Z","lastTransitionTime":"2026-01-26T16:59:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.849905 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.849998 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.850011 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.850034 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.850048 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:36Z","lastTransitionTime":"2026-01-26T16:59:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.952567 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.952919 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.952929 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.952944 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:36 crc kubenswrapper[4856]: I0126 16:59:36.952955 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:36Z","lastTransitionTime":"2026-01-26T16:59:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.055884 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.055924 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.055933 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.055948 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.055959 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:37Z","lastTransitionTime":"2026-01-26T16:59:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.158820 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.158861 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.158872 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.158891 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.158904 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:37Z","lastTransitionTime":"2026-01-26T16:59:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.262186 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.262278 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.262293 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.262319 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.262333 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:37Z","lastTransitionTime":"2026-01-26T16:59:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.365031 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.365078 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.365090 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.365109 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.365123 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:37Z","lastTransitionTime":"2026-01-26T16:59:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.475599 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.475643 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.475658 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.475739 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.475795 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:37Z","lastTransitionTime":"2026-01-26T16:59:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.531928 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 00:58:27.160662964 +0000 UTC Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.578987 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.579013 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.579022 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.579035 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.579044 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:37Z","lastTransitionTime":"2026-01-26T16:59:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.682050 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.682102 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.682113 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.682130 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.682141 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:37Z","lastTransitionTime":"2026-01-26T16:59:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.784431 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.784468 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.784479 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.784496 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.784509 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:37Z","lastTransitionTime":"2026-01-26T16:59:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.886719 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.886790 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.886803 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.886817 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.886827 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:37Z","lastTransitionTime":"2026-01-26T16:59:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.989709 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.989768 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.989789 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.989814 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:37 crc kubenswrapper[4856]: I0126 16:59:37.989830 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:37Z","lastTransitionTime":"2026-01-26T16:59:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.093043 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.093090 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.093103 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.093121 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.093132 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:38Z","lastTransitionTime":"2026-01-26T16:59:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.196406 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.196472 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.196494 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.196519 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.196589 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:38Z","lastTransitionTime":"2026-01-26T16:59:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.299859 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.299920 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.299932 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.299954 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.299968 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:38Z","lastTransitionTime":"2026-01-26T16:59:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.395223 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.395367 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.395409 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:59:38 crc kubenswrapper[4856]: E0126 16:59:38.395443 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.395363 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:59:38 crc kubenswrapper[4856]: E0126 16:59:38.395541 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 16:59:38 crc kubenswrapper[4856]: E0126 16:59:38.395636 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:59:38 crc kubenswrapper[4856]: E0126 16:59:38.395777 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.402673 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.402708 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.402720 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.402735 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.402745 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:38Z","lastTransitionTime":"2026-01-26T16:59:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.506355 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.506433 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.506443 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.506478 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.506490 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:38Z","lastTransitionTime":"2026-01-26T16:59:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.532744 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 20:10:57.690258276 +0000 UTC Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.608948 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.609000 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.609011 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.609032 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.609045 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:38Z","lastTransitionTime":"2026-01-26T16:59:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.712461 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.712540 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.712551 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.712587 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.712603 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:38Z","lastTransitionTime":"2026-01-26T16:59:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.815114 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.815182 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.815194 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.815214 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.815227 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:38Z","lastTransitionTime":"2026-01-26T16:59:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.923370 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.923428 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.923445 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.923465 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:38 crc kubenswrapper[4856]: I0126 16:59:38.923478 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:38Z","lastTransitionTime":"2026-01-26T16:59:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.027189 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.027270 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.027287 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.027310 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.027325 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:39Z","lastTransitionTime":"2026-01-26T16:59:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.098024 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.098090 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.098107 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.098131 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.098148 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:39Z","lastTransitionTime":"2026-01-26T16:59:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:39 crc kubenswrapper[4856]: E0126 16:59:39.118421 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17523591-a778-4a97-aeab-8a7a93101850\\\",\\\"systemUUID\\\":\\\"ca45d056-99cb-4442-8a44-7e899628ecb2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:39Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.122854 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.122898 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.122909 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.122926 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.122937 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:39Z","lastTransitionTime":"2026-01-26T16:59:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:39 crc kubenswrapper[4856]: E0126 16:59:39.136640 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17523591-a778-4a97-aeab-8a7a93101850\\\",\\\"systemUUID\\\":\\\"ca45d056-99cb-4442-8a44-7e899628ecb2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:39Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.140911 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.140944 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.140952 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.140964 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.140973 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:39Z","lastTransitionTime":"2026-01-26T16:59:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:39 crc kubenswrapper[4856]: E0126 16:59:39.165702 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17523591-a778-4a97-aeab-8a7a93101850\\\",\\\"systemUUID\\\":\\\"ca45d056-99cb-4442-8a44-7e899628ecb2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:39Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.170074 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.170107 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.170115 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.170128 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.170138 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:39Z","lastTransitionTime":"2026-01-26T16:59:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:39 crc kubenswrapper[4856]: E0126 16:59:39.184439 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17523591-a778-4a97-aeab-8a7a93101850\\\",\\\"systemUUID\\\":\\\"ca45d056-99cb-4442-8a44-7e899628ecb2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:39Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.189301 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.189340 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.189350 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.189367 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.189377 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:39Z","lastTransitionTime":"2026-01-26T16:59:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:39 crc kubenswrapper[4856]: E0126 16:59:39.201001 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17523591-a778-4a97-aeab-8a7a93101850\\\",\\\"systemUUID\\\":\\\"ca45d056-99cb-4442-8a44-7e899628ecb2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:39Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:39 crc kubenswrapper[4856]: E0126 16:59:39.201145 4856 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.202565 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.202587 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.202594 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.202611 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.202622 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:39Z","lastTransitionTime":"2026-01-26T16:59:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.304870 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.304906 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.304914 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.304927 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.304937 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:39Z","lastTransitionTime":"2026-01-26T16:59:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.407774 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.407822 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.407839 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.407864 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.407881 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:39Z","lastTransitionTime":"2026-01-26T16:59:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.511054 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.511140 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.511158 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.511183 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.511202 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:39Z","lastTransitionTime":"2026-01-26T16:59:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.532869 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 11:40:46.268132806 +0000 UTC Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.614189 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.614246 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.614260 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.614281 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.614298 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:39Z","lastTransitionTime":"2026-01-26T16:59:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.717565 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.717614 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.717625 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.717646 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.717659 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:39Z","lastTransitionTime":"2026-01-26T16:59:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.820188 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.820417 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.820439 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.820466 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.820487 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:39Z","lastTransitionTime":"2026-01-26T16:59:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.925915 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.925968 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.925984 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.926007 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:39 crc kubenswrapper[4856]: I0126 16:59:39.926023 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:39Z","lastTransitionTime":"2026-01-26T16:59:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.028884 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.028947 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.028970 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.028999 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.029020 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:40Z","lastTransitionTime":"2026-01-26T16:59:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.132515 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.132576 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.132588 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.132605 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.132618 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:40Z","lastTransitionTime":"2026-01-26T16:59:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.236119 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.236189 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.236222 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.236251 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.236273 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:40Z","lastTransitionTime":"2026-01-26T16:59:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.339220 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.339287 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.339304 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.339331 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.339354 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:40Z","lastTransitionTime":"2026-01-26T16:59:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.395041 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.395143 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:59:40 crc kubenswrapper[4856]: E0126 16:59:40.395178 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.395429 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 16:59:40 crc kubenswrapper[4856]: E0126 16:59:40.395445 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.395564 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:59:40 crc kubenswrapper[4856]: E0126 16:59:40.395691 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 16:59:40 crc kubenswrapper[4856]: E0126 16:59:40.395950 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.441719 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.441755 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.441767 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.441785 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.441799 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:40Z","lastTransitionTime":"2026-01-26T16:59:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.533328 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 03:26:01.176271514 +0000 UTC Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.544155 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.544190 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.544202 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.544218 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.544229 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:40Z","lastTransitionTime":"2026-01-26T16:59:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.647308 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.647356 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.647371 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.647393 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.647410 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:40Z","lastTransitionTime":"2026-01-26T16:59:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.749092 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.749130 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.749161 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.749178 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.749189 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:40Z","lastTransitionTime":"2026-01-26T16:59:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.852644 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.852710 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.852746 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.852766 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.852778 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:40Z","lastTransitionTime":"2026-01-26T16:59:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.956058 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.956149 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.956185 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.956216 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:40 crc kubenswrapper[4856]: I0126 16:59:40.956237 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:40Z","lastTransitionTime":"2026-01-26T16:59:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.058950 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.058986 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.058997 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.059012 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.059023 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:41Z","lastTransitionTime":"2026-01-26T16:59:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.160596 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.160635 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.160644 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.160657 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.160666 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:41Z","lastTransitionTime":"2026-01-26T16:59:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.263667 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.263707 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.263753 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.263771 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.263783 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:41Z","lastTransitionTime":"2026-01-26T16:59:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.365873 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.365922 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.365936 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.365956 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.365970 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:41Z","lastTransitionTime":"2026-01-26T16:59:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.468813 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.468858 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.468875 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.468891 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.468902 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:41Z","lastTransitionTime":"2026-01-26T16:59:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.533993 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 23:13:54.894630923 +0000 UTC Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.571614 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.571666 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.571681 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.571702 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.571718 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:41Z","lastTransitionTime":"2026-01-26T16:59:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.674600 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.674660 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.674679 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.674703 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.674718 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:41Z","lastTransitionTime":"2026-01-26T16:59:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.777808 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.777858 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.777885 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.777908 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.777922 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:41Z","lastTransitionTime":"2026-01-26T16:59:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.881967 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.882018 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.882063 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.882098 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.882122 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:41Z","lastTransitionTime":"2026-01-26T16:59:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.987699 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.987756 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.987778 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.987812 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:41 crc kubenswrapper[4856]: I0126 16:59:41.987837 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:41Z","lastTransitionTime":"2026-01-26T16:59:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.091292 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.091361 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.091405 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.091437 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.091461 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:42Z","lastTransitionTime":"2026-01-26T16:59:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.197508 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.197600 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.197627 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.197653 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.197670 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:42Z","lastTransitionTime":"2026-01-26T16:59:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.285638 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12e50462-28e6-4531-ada4-e652310e6cce-metrics-certs\") pod \"network-metrics-daemon-295wr\" (UID: \"12e50462-28e6-4531-ada4-e652310e6cce\") " pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 16:59:42 crc kubenswrapper[4856]: E0126 16:59:42.285818 4856 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 16:59:42 crc kubenswrapper[4856]: E0126 16:59:42.285891 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12e50462-28e6-4531-ada4-e652310e6cce-metrics-certs podName:12e50462-28e6-4531-ada4-e652310e6cce nodeName:}" failed. No retries permitted until 2026-01-26 17:00:14.285864872 +0000 UTC m=+110.239118853 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/12e50462-28e6-4531-ada4-e652310e6cce-metrics-certs") pod "network-metrics-daemon-295wr" (UID: "12e50462-28e6-4531-ada4-e652310e6cce") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.300512 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.300593 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.300612 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.300635 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.300652 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:42Z","lastTransitionTime":"2026-01-26T16:59:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.394938 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.394989 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.394965 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.394948 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 16:59:42 crc kubenswrapper[4856]: E0126 16:59:42.395107 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:59:42 crc kubenswrapper[4856]: E0126 16:59:42.395270 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:59:42 crc kubenswrapper[4856]: E0126 16:59:42.395306 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 16:59:42 crc kubenswrapper[4856]: E0126 16:59:42.395347 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.403962 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.403994 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.404003 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.404018 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.404032 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:42Z","lastTransitionTime":"2026-01-26T16:59:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.507007 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.507042 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.507052 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.507068 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.507077 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:42Z","lastTransitionTime":"2026-01-26T16:59:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.534778 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 21:33:37.338164612 +0000 UTC Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.609821 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.609887 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.609905 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.609930 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.609949 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:42Z","lastTransitionTime":"2026-01-26T16:59:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.712839 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.712907 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.712930 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.712959 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.712980 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:42Z","lastTransitionTime":"2026-01-26T16:59:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.816126 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.816202 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.816221 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.816255 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.816290 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:42Z","lastTransitionTime":"2026-01-26T16:59:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.919670 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.919744 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.919765 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.919793 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:42 crc kubenswrapper[4856]: I0126 16:59:42.919819 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:42Z","lastTransitionTime":"2026-01-26T16:59:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.022823 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.022861 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.022873 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.022890 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.022903 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:43Z","lastTransitionTime":"2026-01-26T16:59:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.125408 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.125467 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.125484 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.125513 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.125556 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:43Z","lastTransitionTime":"2026-01-26T16:59:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.228360 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.228402 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.228418 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.228438 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.228453 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:43Z","lastTransitionTime":"2026-01-26T16:59:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.330855 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.330910 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.330921 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.330941 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.330959 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:43Z","lastTransitionTime":"2026-01-26T16:59:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.434139 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.434219 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.434237 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.434262 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.434280 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:43Z","lastTransitionTime":"2026-01-26T16:59:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.535120 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 01:49:19.238651524 +0000 UTC Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.537251 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.537331 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.537359 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.537392 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.537416 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:43Z","lastTransitionTime":"2026-01-26T16:59:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.641298 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.641371 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.641404 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.641435 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.641459 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:43Z","lastTransitionTime":"2026-01-26T16:59:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.744682 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.744736 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.744752 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.744782 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.744798 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:43Z","lastTransitionTime":"2026-01-26T16:59:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.848623 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.848695 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.848719 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.848749 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.848770 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:43Z","lastTransitionTime":"2026-01-26T16:59:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.951777 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.951859 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.951880 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.951908 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:43 crc kubenswrapper[4856]: I0126 16:59:43.951926 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:43Z","lastTransitionTime":"2026-01-26T16:59:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.055655 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.055703 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.055720 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.055743 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.055760 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:44Z","lastTransitionTime":"2026-01-26T16:59:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.158650 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.158713 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.158732 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.158758 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.158776 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:44Z","lastTransitionTime":"2026-01-26T16:59:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.262265 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.262372 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.263852 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.264010 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.264704 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:44Z","lastTransitionTime":"2026-01-26T16:59:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.367886 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.367914 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.367924 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.367937 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.367945 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:44Z","lastTransitionTime":"2026-01-26T16:59:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.394788 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:59:44 crc kubenswrapper[4856]: E0126 16:59:44.394950 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.395698 4856 scope.go:117] "RemoveContainer" containerID="71cb66e6f52823e7aa63f88ba1d153fde73816120aa75d4a6b910937303d2b9e" Jan 26 16:59:44 crc kubenswrapper[4856]: E0126 16:59:44.395826 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pxh94_openshift-ovn-kubernetes(ab5b6f50-172b-4535-a0f9-5d103bcab4e7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.395925 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:59:44 crc kubenswrapper[4856]: E0126 16:59:44.395970 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.396063 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:59:44 crc kubenswrapper[4856]: E0126 16:59:44.396103 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.396191 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 16:59:44 crc kubenswrapper[4856]: E0126 16:59:44.396238 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.470283 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.470338 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.470350 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.470366 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.470376 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:44Z","lastTransitionTime":"2026-01-26T16:59:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.536252 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 16:35:26.149214412 +0000 UTC Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.573948 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.574002 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.574017 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.574036 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.574049 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:44Z","lastTransitionTime":"2026-01-26T16:59:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.677036 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.677108 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.677133 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.677166 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.677189 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:44Z","lastTransitionTime":"2026-01-26T16:59:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.779592 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.779847 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.779885 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.779942 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.779967 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:44Z","lastTransitionTime":"2026-01-26T16:59:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.883602 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.883652 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.883669 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.883694 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.883711 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:44Z","lastTransitionTime":"2026-01-26T16:59:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.986425 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.986502 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.986521 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.986580 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:44 crc kubenswrapper[4856]: I0126 16:59:44.986598 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:44Z","lastTransitionTime":"2026-01-26T16:59:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.089715 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.089830 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.089850 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.089875 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.089893 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:45Z","lastTransitionTime":"2026-01-26T16:59:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.193600 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.193644 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.193656 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.193676 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.193690 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:45Z","lastTransitionTime":"2026-01-26T16:59:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.297573 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.297623 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.297634 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.297657 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.297669 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:45Z","lastTransitionTime":"2026-01-26T16:59:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.401170 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.401217 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.401229 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.401246 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.401257 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:45Z","lastTransitionTime":"2026-01-26T16:59:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.409199 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:45Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.410374 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.425953 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad7b59f9-beb7-49d6-a2d1-e29133e46854\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53a6a6d4222ee5e1ac29fd5957d8cdf3fd42de72c68c85329374ce7afc4004\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v2l7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:45Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.446909 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71cb66e6f52823e7aa63f88ba1d153fde73816120aa75d4a6b910937303d2b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71cb66e6f52823e7aa63f88ba1d153fde73816120aa75d4a6b910937303d2b9e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:59:29Z\\\",\\\"message\\\":\\\"n-kubernetes/ovnkube-node-pxh94\\\\nI0126 16:59:29.834990 6550 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-t4fq2\\\\nI0126 16:59:29.834996 6550 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-t4fq2\\\\nI0126 16:59:29.835005 6550 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-t4fq2 in node crc\\\\nI0126 16:59:29.835026 6550 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0126 16:59:29.835056 6550 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0126 16:59:29.835079 6550 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0126 16:59:29.835090 6550 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nF0126 16:59:29.835097 6550 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pxh94_openshift-ovn-kubernetes(ab5b6f50-172b-4535-a0f9-5d103bcab4e7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pxh94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:45Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.461094 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-295wr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12e50462-28e6-4531-ada4-e652310e6cce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-295wr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:45Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.476012 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc864b0d-83bc-4954-9c61-ad650157caff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbde934a6c8acad10ca3ab8206d0ddbd4f7b17e9d304b898a68f4d3b0303bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b48f763d4aff37169399be766d5ab4f7ebbf91f304d139c9022a8556946eb107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0094e662f53c4832a984e05a880021af05ffc4c27f25394c28a070d9ef5490d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb95f623fa4f5f42217649ccce4225f9fe588bb3558da9334eef2b40ddb62486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb95f623fa4f5f42217649ccce4225f9fe588bb3558da9334eef2b40ddb62486\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:45Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.487578 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63c75ede-5170-4db0-811b-5217ef8d72b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da26ddcfc6ccde3c9aabd63bdef3435f9a9eaab8c095bfeef5670c1295576cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54ca3fa13d9e8d442efa93b44a870369f7df3fe7562d77b98528f5c19a751f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xm9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:45Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.500038 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v7579" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a77e85f9-b566-4807-bb92-55963c97b93c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba87c9fc35c230bbee201a5176cb467309f0b9aee82dfc81f3b677a15486d02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c03dc794e9c2035f2e1983eacad3e51d76223cb1b82e2f402c73f9453e4bd2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v7579\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:45Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.503739 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.503770 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.503779 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.503793 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.503803 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:45Z","lastTransitionTime":"2026-01-26T16:59:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.512050 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af4930ff9cebd62106545fb3da2dfc93f7b591426fe4c85aa2e637b60c935f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:45Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.524127 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c09748c0963bdb50c9552f249d1135ea53e68c52babd788dad050951cb849cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:45Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.536844 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 18:48:04.984716375 +0000 UTC Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.539308 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:45Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.551340 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:45Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.561623 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://627cd6b39fdedd967fffcfc0755439277b2d73016fba1ddc7f5b9d5deba43b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:45Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.576160 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rq622" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a742e7b-c420-46e3-9e96-e9c744af6124\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad7222c9b91c0065a545bc1904d9864a5923cc13bfb6617daeb4a965a830f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8plh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rq622\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:45Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.589764 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ecd87a-c5db-446d-ad3e-cfabbd648c1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89975b8f9428f81ab5d3fb48ced5dd9c837bea2feea3b89f5f7ff8d7d5d15b3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d432159a07ba20bf95f058bf8a597d67cbd8d852519bed035694f8ba3d8ec4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59ba0f131eee14df389391e46bc772894e49e976c3663df5fb3bc98b3cb4a3d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03a01b651d9f66da4dd1f6e0d29ad97c0d6ae46b644c3d997d8dc99476706df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f07438e20bdf71c752bb661084c835341999c561bb5442d75e177223881276f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:58:51Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:58:43.431198 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:58:43.432227 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99450334/tls.crt::/tmp/serving-cert-99450334/tls.key\\\\\\\"\\\\nI0126 16:58:50.828320 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:58:50.830449 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:58:50.830472 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:58:50.830509 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:58:50.830515 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:58:50.834885 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:58:50.834958 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:58:50.834991 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0126 16:58:50.834905 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0126 16:58:50.835015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:58:50.835084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:58:50.835106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:58:50.835116 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0126 16:58:50.837641 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2daa3f13d37b7ae19dfca406b6b5cfdcd4f211287f7d410193f2fee36a24553\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:45Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.601836 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191ad5d41024c88c1a4fbc30f307dd6340fd55b93b00d22f62209d0e82be286f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2a4d8be409a46ffe5702797e56a80023a72a19fb7dc5c49e5f4984cedc600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:45Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.606380 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.606410 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.606420 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.606435 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.606446 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:45Z","lastTransitionTime":"2026-01-26T16:59:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.626234 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c823748751e9938f10fab08e33b5fffff5a6d15961ce028204131b7b69a56c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:45Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.708503 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.708557 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.708569 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.708587 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.708602 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:45Z","lastTransitionTime":"2026-01-26T16:59:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.811443 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.811499 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.811513 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.811554 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.811571 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:45Z","lastTransitionTime":"2026-01-26T16:59:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.914398 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.914440 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.914449 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.914466 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:45 crc kubenswrapper[4856]: I0126 16:59:45.914478 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:45Z","lastTransitionTime":"2026-01-26T16:59:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.017762 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.017792 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.017801 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.017814 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.017823 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:46Z","lastTransitionTime":"2026-01-26T16:59:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.120154 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.120194 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.120205 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.120221 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.120232 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:46Z","lastTransitionTime":"2026-01-26T16:59:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.223305 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.223340 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.223352 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.223371 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.223383 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:46Z","lastTransitionTime":"2026-01-26T16:59:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.326750 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.326780 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.326791 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.326805 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.326814 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:46Z","lastTransitionTime":"2026-01-26T16:59:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.394792 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.394892 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:59:46 crc kubenswrapper[4856]: E0126 16:59:46.394969 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:59:46 crc kubenswrapper[4856]: E0126 16:59:46.395018 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.395079 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:59:46 crc kubenswrapper[4856]: E0126 16:59:46.395122 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.395142 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 16:59:46 crc kubenswrapper[4856]: E0126 16:59:46.395187 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.430122 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.430155 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.430163 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.430180 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.430188 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:46Z","lastTransitionTime":"2026-01-26T16:59:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.533924 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.534014 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.534040 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.534072 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.534097 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:46Z","lastTransitionTime":"2026-01-26T16:59:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.537607 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 02:34:08.921743991 +0000 UTC Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.637216 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.637268 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.637278 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.637294 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.637303 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:46Z","lastTransitionTime":"2026-01-26T16:59:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.740178 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.740246 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.740266 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.740295 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.740312 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:46Z","lastTransitionTime":"2026-01-26T16:59:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.803753 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-rq622_7a742e7b-c420-46e3-9e96-e9c744af6124/kube-multus/0.log" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.803813 4856 generic.go:334] "Generic (PLEG): container finished" podID="7a742e7b-c420-46e3-9e96-e9c744af6124" containerID="ad7222c9b91c0065a545bc1904d9864a5923cc13bfb6617daeb4a965a830f191" exitCode=1 Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.803852 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-rq622" event={"ID":"7a742e7b-c420-46e3-9e96-e9c744af6124","Type":"ContainerDied","Data":"ad7222c9b91c0065a545bc1904d9864a5923cc13bfb6617daeb4a965a830f191"} Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.804274 4856 scope.go:117] "RemoveContainer" containerID="ad7222c9b91c0065a545bc1904d9864a5923cc13bfb6617daeb4a965a830f191" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.830273 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:46Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.843817 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.843870 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.843884 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.843906 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.843923 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:46Z","lastTransitionTime":"2026-01-26T16:59:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.851926 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad7b59f9-beb7-49d6-a2d1-e29133e46854\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53a6a6d4222ee5e1ac29fd5957d8cdf3fd42de72c68c85329374ce7afc4004\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v2l7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:46Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.889034 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71cb66e6f52823e7aa63f88ba1d153fde73816120aa75d4a6b910937303d2b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71cb66e6f52823e7aa63f88ba1d153fde73816120aa75d4a6b910937303d2b9e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:59:29Z\\\",\\\"message\\\":\\\"n-kubernetes/ovnkube-node-pxh94\\\\nI0126 16:59:29.834990 6550 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-t4fq2\\\\nI0126 16:59:29.834996 6550 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-t4fq2\\\\nI0126 16:59:29.835005 6550 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-t4fq2 in node crc\\\\nI0126 16:59:29.835026 6550 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0126 16:59:29.835056 6550 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0126 16:59:29.835079 6550 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0126 16:59:29.835090 6550 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nF0126 16:59:29.835097 6550 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pxh94_openshift-ovn-kubernetes(ab5b6f50-172b-4535-a0f9-5d103bcab4e7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pxh94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:46Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.909881 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-295wr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12e50462-28e6-4531-ada4-e652310e6cce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-295wr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:46Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.930977 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc864b0d-83bc-4954-9c61-ad650157caff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbde934a6c8acad10ca3ab8206d0ddbd4f7b17e9d304b898a68f4d3b0303bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b48f763d4aff37169399be766d5ab4f7ebbf91f304d139c9022a8556946eb107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0094e662f53c4832a984e05a880021af05ffc4c27f25394c28a070d9ef5490d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb95f623fa4f5f42217649ccce4225f9fe588bb3558da9334eef2b40ddb62486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb95f623fa4f5f42217649ccce4225f9fe588bb3558da9334eef2b40ddb62486\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:46Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.947052 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c3b0574-b4cc-483d-ae88-6517d1f30772\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec9063a7c03990fc26fc47427f164a769fd649c2bdbd9d23ea7f646e569734be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67797295a8c3952902b2696c6fdb26b72ce1826b5ccd522a24aac90a0411b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f67797295a8c3952902b2696c6fdb26b72ce1826b5ccd522a24aac90a0411b5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:46Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.950478 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.950511 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.950537 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.950557 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.950569 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:46Z","lastTransitionTime":"2026-01-26T16:59:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.965690 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63c75ede-5170-4db0-811b-5217ef8d72b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da26ddcfc6ccde3c9aabd63bdef3435f9a9eaab8c095bfeef5670c1295576cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54ca3fa13d9e8d442efa93b44a870369f7df3fe7562d77b98528f5c19a751f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xm9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:46Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.983351 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v7579" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a77e85f9-b566-4807-bb92-55963c97b93c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba87c9fc35c230bbee201a5176cb467309f0b9aee82dfc81f3b677a15486d02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c03dc794e9c2035f2e1983eacad3e51d76223cb1b82e2f402c73f9453e4bd2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v7579\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:46Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:46 crc kubenswrapper[4856]: I0126 16:59:46.997983 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af4930ff9cebd62106545fb3da2dfc93f7b591426fe4c85aa2e637b60c935f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:46Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.016810 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c09748c0963bdb50c9552f249d1135ea53e68c52babd788dad050951cb849cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.031336 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.046932 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.052800 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.052840 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.052849 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.052863 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.052872 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:47Z","lastTransitionTime":"2026-01-26T16:59:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.062040 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://627cd6b39fdedd967fffcfc0755439277b2d73016fba1ddc7f5b9d5deba43b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.077063 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rq622" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a742e7b-c420-46e3-9e96-e9c744af6124\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad7222c9b91c0065a545bc1904d9864a5923cc13bfb6617daeb4a965a830f191\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad7222c9b91c0065a545bc1904d9864a5923cc13bfb6617daeb4a965a830f191\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:59:45Z\\\",\\\"message\\\":\\\"2026-01-26T16:59:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_41ab0694-d9c8-49a7-bf30-57e732ac7550\\\\n2026-01-26T16:59:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_41ab0694-d9c8-49a7-bf30-57e732ac7550 to /host/opt/cni/bin/\\\\n2026-01-26T16:59:00Z [verbose] multus-daemon started\\\\n2026-01-26T16:59:00Z [verbose] Readiness Indicator file check\\\\n2026-01-26T16:59:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8plh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rq622\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.092589 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ecd87a-c5db-446d-ad3e-cfabbd648c1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89975b8f9428f81ab5d3fb48ced5dd9c837bea2feea3b89f5f7ff8d7d5d15b3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d432159a07ba20bf95f058bf8a597d67cbd8d852519bed035694f8ba3d8ec4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59ba0f131eee14df389391e46bc772894e49e976c3663df5fb3bc98b3cb4a3d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03a01b651d9f66da4dd1f6e0d29ad97c0d6ae46b644c3d997d8dc99476706df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f07438e20bdf71c752bb661084c835341999c561bb5442d75e177223881276f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:58:51Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:58:43.431198 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:58:43.432227 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99450334/tls.crt::/tmp/serving-cert-99450334/tls.key\\\\\\\"\\\\nI0126 16:58:50.828320 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:58:50.830449 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:58:50.830472 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:58:50.830509 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:58:50.830515 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:58:50.834885 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:58:50.834958 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:58:50.834991 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0126 16:58:50.834905 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0126 16:58:50.835015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:58:50.835084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:58:50.835106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:58:50.835116 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0126 16:58:50.837641 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2daa3f13d37b7ae19dfca406b6b5cfdcd4f211287f7d410193f2fee36a24553\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.106157 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191ad5d41024c88c1a4fbc30f307dd6340fd55b93b00d22f62209d0e82be286f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2a4d8be409a46ffe5702797e56a80023a72a19fb7dc5c49e5f4984cedc600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.119922 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c823748751e9938f10fab08e33b5fffff5a6d15961ce028204131b7b69a56c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.155154 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.155190 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.155201 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.155218 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.155229 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:47Z","lastTransitionTime":"2026-01-26T16:59:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.258300 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.258572 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.258641 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.258708 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.258770 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:47Z","lastTransitionTime":"2026-01-26T16:59:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.362163 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.362209 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.362221 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.362238 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.362251 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:47Z","lastTransitionTime":"2026-01-26T16:59:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.465295 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.465361 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.465378 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.465401 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.465419 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:47Z","lastTransitionTime":"2026-01-26T16:59:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.538384 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 18:20:50.526085108 +0000 UTC Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.570169 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.570355 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.570456 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.570628 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.570765 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:47Z","lastTransitionTime":"2026-01-26T16:59:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.674571 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.674942 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.675099 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.675241 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.675378 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:47Z","lastTransitionTime":"2026-01-26T16:59:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.778176 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.778459 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.778586 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.778669 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.778734 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:47Z","lastTransitionTime":"2026-01-26T16:59:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.811148 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-rq622_7a742e7b-c420-46e3-9e96-e9c744af6124/kube-multus/0.log" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.811429 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-rq622" event={"ID":"7a742e7b-c420-46e3-9e96-e9c744af6124","Type":"ContainerStarted","Data":"afeb20035224feeab28a92ac77b43a24e653e49c56a25590a9861019a2b7a8ff"} Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.827078 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c3b0574-b4cc-483d-ae88-6517d1f30772\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec9063a7c03990fc26fc47427f164a769fd649c2bdbd9d23ea7f646e569734be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67797295a8c3952902b2696c6fdb26b72ce1826b5ccd522a24aac90a0411b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f67797295a8c3952902b2696c6fdb26b72ce1826b5ccd522a24aac90a0411b5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.840718 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63c75ede-5170-4db0-811b-5217ef8d72b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da26ddcfc6ccde3c9aabd63bdef3435f9a9eaab8c095bfeef5670c1295576cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54ca3fa13d9e8d442efa93b44a870369f7df3fe7562d77b98528f5c19a751f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xm9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.856328 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc864b0d-83bc-4954-9c61-ad650157caff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbde934a6c8acad10ca3ab8206d0ddbd4f7b17e9d304b898a68f4d3b0303bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b48f763d4aff37169399be766d5ab4f7ebbf91f304d139c9022a8556946eb107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0094e662f53c4832a984e05a880021af05ffc4c27f25394c28a070d9ef5490d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb95f623fa4f5f42217649ccce4225f9fe588bb3558da9334eef2b40ddb62486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb95f623fa4f5f42217649ccce4225f9fe588bb3558da9334eef2b40ddb62486\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.874101 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af4930ff9cebd62106545fb3da2dfc93f7b591426fe4c85aa2e637b60c935f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.881754 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.881789 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.881799 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.881812 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.881821 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:47Z","lastTransitionTime":"2026-01-26T16:59:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.891371 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c09748c0963bdb50c9552f249d1135ea53e68c52babd788dad050951cb849cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.909929 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.929032 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v7579" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a77e85f9-b566-4807-bb92-55963c97b93c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba87c9fc35c230bbee201a5176cb467309f0b9aee82dfc81f3b677a15486d02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c03dc794e9c2035f2e1983eacad3e51d76223cb1b82e2f402c73f9453e4bd2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v7579\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.946342 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191ad5d41024c88c1a4fbc30f307dd6340fd55b93b00d22f62209d0e82be286f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2a4d8be409a46ffe5702797e56a80023a72a19fb7dc5c49e5f4984cedc600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.963866 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c823748751e9938f10fab08e33b5fffff5a6d15961ce028204131b7b69a56c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.978248 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.984670 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.984719 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.984737 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.984761 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.984778 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:47Z","lastTransitionTime":"2026-01-26T16:59:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:47 crc kubenswrapper[4856]: I0126 16:59:47.995424 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://627cd6b39fdedd967fffcfc0755439277b2d73016fba1ddc7f5b9d5deba43b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:47Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.015012 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rq622" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a742e7b-c420-46e3-9e96-e9c744af6124\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afeb20035224feeab28a92ac77b43a24e653e49c56a25590a9861019a2b7a8ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad7222c9b91c0065a545bc1904d9864a5923cc13bfb6617daeb4a965a830f191\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:59:45Z\\\",\\\"message\\\":\\\"2026-01-26T16:59:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_41ab0694-d9c8-49a7-bf30-57e732ac7550\\\\n2026-01-26T16:59:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_41ab0694-d9c8-49a7-bf30-57e732ac7550 to /host/opt/cni/bin/\\\\n2026-01-26T16:59:00Z [verbose] multus-daemon started\\\\n2026-01-26T16:59:00Z [verbose] Readiness Indicator file check\\\\n2026-01-26T16:59:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8plh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rq622\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:48Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.038355 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ecd87a-c5db-446d-ad3e-cfabbd648c1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89975b8f9428f81ab5d3fb48ced5dd9c837bea2feea3b89f5f7ff8d7d5d15b3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d432159a07ba20bf95f058bf8a597d67cbd8d852519bed035694f8ba3d8ec4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59ba0f131eee14df389391e46bc772894e49e976c3663df5fb3bc98b3cb4a3d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03a01b651d9f66da4dd1f6e0d29ad97c0d6ae46b644c3d997d8dc99476706df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f07438e20bdf71c752bb661084c835341999c561bb5442d75e177223881276f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:58:51Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:58:43.431198 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:58:43.432227 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99450334/tls.crt::/tmp/serving-cert-99450334/tls.key\\\\\\\"\\\\nI0126 16:58:50.828320 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:58:50.830449 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:58:50.830472 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:58:50.830509 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:58:50.830515 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:58:50.834885 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:58:50.834958 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:58:50.834991 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0126 16:58:50.834905 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0126 16:58:50.835015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:58:50.835084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:58:50.835106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:58:50.835116 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0126 16:58:50.837641 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2daa3f13d37b7ae19dfca406b6b5cfdcd4f211287f7d410193f2fee36a24553\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:48Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.060392 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad7b59f9-beb7-49d6-a2d1-e29133e46854\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53a6a6d4222ee5e1ac29fd5957d8cdf3fd42de72c68c85329374ce7afc4004\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v2l7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:48Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.087563 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.087641 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.087662 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.087694 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.087718 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:48Z","lastTransitionTime":"2026-01-26T16:59:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.096818 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71cb66e6f52823e7aa63f88ba1d153fde73816120aa75d4a6b910937303d2b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71cb66e6f52823e7aa63f88ba1d153fde73816120aa75d4a6b910937303d2b9e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:59:29Z\\\",\\\"message\\\":\\\"n-kubernetes/ovnkube-node-pxh94\\\\nI0126 16:59:29.834990 6550 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-t4fq2\\\\nI0126 16:59:29.834996 6550 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-t4fq2\\\\nI0126 16:59:29.835005 6550 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-t4fq2 in node crc\\\\nI0126 16:59:29.835026 6550 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0126 16:59:29.835056 6550 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0126 16:59:29.835079 6550 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0126 16:59:29.835090 6550 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nF0126 16:59:29.835097 6550 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pxh94_openshift-ovn-kubernetes(ab5b6f50-172b-4535-a0f9-5d103bcab4e7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pxh94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:48Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.113312 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-295wr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12e50462-28e6-4531-ada4-e652310e6cce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-295wr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:48Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.136934 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:48Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.190325 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.190360 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.190372 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.190388 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.190400 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:48Z","lastTransitionTime":"2026-01-26T16:59:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.293382 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.293490 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.293517 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.293915 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.293948 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:48Z","lastTransitionTime":"2026-01-26T16:59:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.395059 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.395125 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.395101 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.395325 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 16:59:48 crc kubenswrapper[4856]: E0126 16:59:48.395648 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:59:48 crc kubenswrapper[4856]: E0126 16:59:48.395766 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:59:48 crc kubenswrapper[4856]: E0126 16:59:48.395913 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:59:48 crc kubenswrapper[4856]: E0126 16:59:48.396059 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.396812 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.396901 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.396922 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.396948 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.396966 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:48Z","lastTransitionTime":"2026-01-26T16:59:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.499676 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.499730 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.499746 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.499769 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.499787 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:48Z","lastTransitionTime":"2026-01-26T16:59:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.539953 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 09:01:55.747805651 +0000 UTC Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.602948 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.603018 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.603039 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.603065 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.603088 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:48Z","lastTransitionTime":"2026-01-26T16:59:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.719500 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.719632 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.719650 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.719671 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.719685 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:48Z","lastTransitionTime":"2026-01-26T16:59:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.822317 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.822370 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.822383 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.822408 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.822424 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:48Z","lastTransitionTime":"2026-01-26T16:59:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.925613 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.925662 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.925677 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.925698 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:48 crc kubenswrapper[4856]: I0126 16:59:48.925728 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:48Z","lastTransitionTime":"2026-01-26T16:59:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.028163 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.028254 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.028288 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.028320 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.028344 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:49Z","lastTransitionTime":"2026-01-26T16:59:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.131290 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.131347 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.131358 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.131383 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.131395 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:49Z","lastTransitionTime":"2026-01-26T16:59:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.234172 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.234219 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.234231 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.234247 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.234260 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:49Z","lastTransitionTime":"2026-01-26T16:59:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.337935 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.337983 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.338000 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.338022 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.338037 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:49Z","lastTransitionTime":"2026-01-26T16:59:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.441326 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.441424 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.441479 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.441521 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.441595 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:49Z","lastTransitionTime":"2026-01-26T16:59:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.540190 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 21:26:29.075052447 +0000 UTC Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.544998 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.545053 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.545067 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.545085 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.545097 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:49Z","lastTransitionTime":"2026-01-26T16:59:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.565325 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.565367 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.565378 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.565393 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.565404 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:49Z","lastTransitionTime":"2026-01-26T16:59:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:49 crc kubenswrapper[4856]: E0126 16:59:49.579126 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17523591-a778-4a97-aeab-8a7a93101850\\\",\\\"systemUUID\\\":\\\"ca45d056-99cb-4442-8a44-7e899628ecb2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:49Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.583552 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.583603 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.583618 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.583635 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.583644 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:49Z","lastTransitionTime":"2026-01-26T16:59:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:49 crc kubenswrapper[4856]: E0126 16:59:49.597626 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17523591-a778-4a97-aeab-8a7a93101850\\\",\\\"systemUUID\\\":\\\"ca45d056-99cb-4442-8a44-7e899628ecb2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:49Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.602362 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.602411 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.602422 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.602441 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.602453 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:49Z","lastTransitionTime":"2026-01-26T16:59:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:49 crc kubenswrapper[4856]: E0126 16:59:49.617024 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17523591-a778-4a97-aeab-8a7a93101850\\\",\\\"systemUUID\\\":\\\"ca45d056-99cb-4442-8a44-7e899628ecb2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:49Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.621297 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.621339 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.621350 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.621367 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.621380 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:49Z","lastTransitionTime":"2026-01-26T16:59:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:49 crc kubenswrapper[4856]: E0126 16:59:49.636208 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17523591-a778-4a97-aeab-8a7a93101850\\\",\\\"systemUUID\\\":\\\"ca45d056-99cb-4442-8a44-7e899628ecb2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:49Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.640673 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.640705 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.640714 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.640732 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.640744 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:49Z","lastTransitionTime":"2026-01-26T16:59:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:49 crc kubenswrapper[4856]: E0126 16:59:49.652399 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17523591-a778-4a97-aeab-8a7a93101850\\\",\\\"systemUUID\\\":\\\"ca45d056-99cb-4442-8a44-7e899628ecb2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:49Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:49 crc kubenswrapper[4856]: E0126 16:59:49.652550 4856 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.654472 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.654501 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.654509 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.654534 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.654544 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:49Z","lastTransitionTime":"2026-01-26T16:59:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.757258 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.757306 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.757320 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.757338 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.757351 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:49Z","lastTransitionTime":"2026-01-26T16:59:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.894192 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.894253 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.894268 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.894289 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.894304 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:49Z","lastTransitionTime":"2026-01-26T16:59:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.996673 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.996727 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.996738 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.996758 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:49 crc kubenswrapper[4856]: I0126 16:59:49.996770 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:49Z","lastTransitionTime":"2026-01-26T16:59:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.099595 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.099676 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.099696 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.099723 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.099745 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:50Z","lastTransitionTime":"2026-01-26T16:59:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.202433 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.202489 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.202506 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.202548 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.202565 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:50Z","lastTransitionTime":"2026-01-26T16:59:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.305789 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.305842 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.305854 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.305870 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.305882 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:50Z","lastTransitionTime":"2026-01-26T16:59:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.394506 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.394585 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.394630 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.394585 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:59:50 crc kubenswrapper[4856]: E0126 16:59:50.394750 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:59:50 crc kubenswrapper[4856]: E0126 16:59:50.394853 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:59:50 crc kubenswrapper[4856]: E0126 16:59:50.395015 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:59:50 crc kubenswrapper[4856]: E0126 16:59:50.395145 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.409443 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.409494 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.409510 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.409552 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.409567 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:50Z","lastTransitionTime":"2026-01-26T16:59:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.512160 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.512229 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.512254 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.512285 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.512310 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:50Z","lastTransitionTime":"2026-01-26T16:59:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.541233 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 15:51:56.746485422 +0000 UTC Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.615514 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.615605 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.615623 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.615688 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.615710 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:50Z","lastTransitionTime":"2026-01-26T16:59:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.718973 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.719024 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.719039 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.719058 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.719069 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:50Z","lastTransitionTime":"2026-01-26T16:59:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.821126 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.821164 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.821174 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.821190 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.821201 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:50Z","lastTransitionTime":"2026-01-26T16:59:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.923497 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.923579 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.923592 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.923610 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:50 crc kubenswrapper[4856]: I0126 16:59:50.923620 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:50Z","lastTransitionTime":"2026-01-26T16:59:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.026539 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.026580 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.026591 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.026625 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.026638 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:51Z","lastTransitionTime":"2026-01-26T16:59:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.129993 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.130060 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.130077 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.130103 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.130122 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:51Z","lastTransitionTime":"2026-01-26T16:59:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.233520 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.233604 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.233615 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.233639 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.233657 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:51Z","lastTransitionTime":"2026-01-26T16:59:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.336051 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.336120 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.336135 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.336151 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.336168 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:51Z","lastTransitionTime":"2026-01-26T16:59:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.439739 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.439823 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.439847 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.439877 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.439899 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:51Z","lastTransitionTime":"2026-01-26T16:59:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.541355 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 12:15:00.701583198 +0000 UTC Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.542765 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.542811 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.542821 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.542838 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.542848 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:51Z","lastTransitionTime":"2026-01-26T16:59:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.645855 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.645898 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.645906 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.645923 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.645934 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:51Z","lastTransitionTime":"2026-01-26T16:59:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.749018 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.749567 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.749606 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.749625 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.749637 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:51Z","lastTransitionTime":"2026-01-26T16:59:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.851350 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.851400 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.851416 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.851434 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.851450 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:51Z","lastTransitionTime":"2026-01-26T16:59:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.954628 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.954682 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.954701 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.954725 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:51 crc kubenswrapper[4856]: I0126 16:59:51.954744 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:51Z","lastTransitionTime":"2026-01-26T16:59:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.058385 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.058436 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.058451 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.058470 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.058485 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:52Z","lastTransitionTime":"2026-01-26T16:59:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.161595 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.161634 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.161655 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.161683 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.161699 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:52Z","lastTransitionTime":"2026-01-26T16:59:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.265058 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.265092 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.265102 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.265122 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.265136 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:52Z","lastTransitionTime":"2026-01-26T16:59:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.368726 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.368802 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.368814 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.368831 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.368843 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:52Z","lastTransitionTime":"2026-01-26T16:59:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.394161 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.394201 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.394207 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.394161 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:59:52 crc kubenswrapper[4856]: E0126 16:59:52.394305 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:59:52 crc kubenswrapper[4856]: E0126 16:59:52.394397 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:59:52 crc kubenswrapper[4856]: E0126 16:59:52.394474 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:59:52 crc kubenswrapper[4856]: E0126 16:59:52.394593 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.471801 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.471846 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.471856 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.471876 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.471888 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:52Z","lastTransitionTime":"2026-01-26T16:59:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.541715 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 00:19:05.375815469 +0000 UTC Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.574635 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.574668 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.574677 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.574692 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.574701 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:52Z","lastTransitionTime":"2026-01-26T16:59:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.677124 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.677187 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.677206 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.677231 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.677249 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:52Z","lastTransitionTime":"2026-01-26T16:59:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.779919 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.780073 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.780099 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.780163 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.780183 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:52Z","lastTransitionTime":"2026-01-26T16:59:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.883124 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.883164 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.883184 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.883206 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.883225 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:52Z","lastTransitionTime":"2026-01-26T16:59:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.985077 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.985130 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.985165 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.985182 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:52 crc kubenswrapper[4856]: I0126 16:59:52.985196 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:52Z","lastTransitionTime":"2026-01-26T16:59:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.088672 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.088745 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.088767 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.088799 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.088822 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:53Z","lastTransitionTime":"2026-01-26T16:59:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.192217 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.192288 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.192300 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.192322 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.192334 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:53Z","lastTransitionTime":"2026-01-26T16:59:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.295672 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.295755 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.295797 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.295830 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.295853 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:53Z","lastTransitionTime":"2026-01-26T16:59:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.398475 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.398547 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.398558 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.398576 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.398660 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:53Z","lastTransitionTime":"2026-01-26T16:59:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.501235 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.501275 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.501288 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.501304 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.501315 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:53Z","lastTransitionTime":"2026-01-26T16:59:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.543090 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 17:23:29.074884746 +0000 UTC Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.603398 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.603453 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.603472 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.603488 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.603499 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:53Z","lastTransitionTime":"2026-01-26T16:59:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.707198 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.707272 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.707296 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.707331 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.707350 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:53Z","lastTransitionTime":"2026-01-26T16:59:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.810627 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.810683 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.810702 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.810729 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.810747 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:53Z","lastTransitionTime":"2026-01-26T16:59:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.913648 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.913783 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.913810 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.913844 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:53 crc kubenswrapper[4856]: I0126 16:59:53.913869 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:53Z","lastTransitionTime":"2026-01-26T16:59:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.017201 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.017246 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.017257 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.017273 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.017284 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:54Z","lastTransitionTime":"2026-01-26T16:59:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.119943 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.119986 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.119997 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.120012 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.120022 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:54Z","lastTransitionTime":"2026-01-26T16:59:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.222755 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.222783 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.222793 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.222808 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.222817 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:54Z","lastTransitionTime":"2026-01-26T16:59:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.325481 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.325563 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.325576 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.325593 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.325603 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:54Z","lastTransitionTime":"2026-01-26T16:59:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.394368 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.394399 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.394762 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:59:54 crc kubenswrapper[4856]: E0126 16:59:54.394916 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.394978 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 16:59:54 crc kubenswrapper[4856]: E0126 16:59:54.395039 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:59:54 crc kubenswrapper[4856]: E0126 16:59:54.395255 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:59:54 crc kubenswrapper[4856]: E0126 16:59:54.395313 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.428670 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.428745 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.428762 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.428784 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.428800 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:54Z","lastTransitionTime":"2026-01-26T16:59:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.531293 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.531349 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.531383 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.531412 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.531438 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:54Z","lastTransitionTime":"2026-01-26T16:59:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.543593 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 08:41:14.922161812 +0000 UTC Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.634393 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.634477 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.634501 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.634568 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.634597 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:54Z","lastTransitionTime":"2026-01-26T16:59:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.737954 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.738011 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.738020 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.738035 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.738064 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:54Z","lastTransitionTime":"2026-01-26T16:59:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.840448 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.840501 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.840553 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.840573 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.840589 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:54Z","lastTransitionTime":"2026-01-26T16:59:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.944047 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.944107 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.944120 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.944139 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:54 crc kubenswrapper[4856]: I0126 16:59:54.944155 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:54Z","lastTransitionTime":"2026-01-26T16:59:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.047443 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.047572 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.047604 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.047633 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.047652 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:55Z","lastTransitionTime":"2026-01-26T16:59:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.151088 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.151174 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.151199 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.151229 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.151251 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:55Z","lastTransitionTime":"2026-01-26T16:59:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.255169 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.255246 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.255263 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.255286 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.255301 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:55Z","lastTransitionTime":"2026-01-26T16:59:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.359348 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.359416 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.359438 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.359487 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.359513 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:55Z","lastTransitionTime":"2026-01-26T16:59:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.410699 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v7579" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a77e85f9-b566-4807-bb92-55963c97b93c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba87c9fc35c230bbee201a5176cb467309f0b9aee82dfc81f3b677a15486d02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c03dc794e9c2035f2e1983eacad3e51d76223cb1b82e2f402c73f9453e4bd2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v7579\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:55Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.424736 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af4930ff9cebd62106545fb3da2dfc93f7b591426fe4c85aa2e637b60c935f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:55Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.442306 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c09748c0963bdb50c9552f249d1135ea53e68c52babd788dad050951cb849cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:55Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.460933 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:55Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.463243 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.463291 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.463302 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.463320 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.463332 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:55Z","lastTransitionTime":"2026-01-26T16:59:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.474851 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:55Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.487251 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://627cd6b39fdedd967fffcfc0755439277b2d73016fba1ddc7f5b9d5deba43b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:55Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.500473 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rq622" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a742e7b-c420-46e3-9e96-e9c744af6124\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afeb20035224feeab28a92ac77b43a24e653e49c56a25590a9861019a2b7a8ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad7222c9b91c0065a545bc1904d9864a5923cc13bfb6617daeb4a965a830f191\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:59:45Z\\\",\\\"message\\\":\\\"2026-01-26T16:59:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_41ab0694-d9c8-49a7-bf30-57e732ac7550\\\\n2026-01-26T16:59:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_41ab0694-d9c8-49a7-bf30-57e732ac7550 to /host/opt/cni/bin/\\\\n2026-01-26T16:59:00Z [verbose] multus-daemon started\\\\n2026-01-26T16:59:00Z [verbose] Readiness Indicator file check\\\\n2026-01-26T16:59:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8plh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rq622\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:55Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.514439 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ecd87a-c5db-446d-ad3e-cfabbd648c1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89975b8f9428f81ab5d3fb48ced5dd9c837bea2feea3b89f5f7ff8d7d5d15b3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d432159a07ba20bf95f058bf8a597d67cbd8d852519bed035694f8ba3d8ec4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59ba0f131eee14df389391e46bc772894e49e976c3663df5fb3bc98b3cb4a3d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03a01b651d9f66da4dd1f6e0d29ad97c0d6ae46b644c3d997d8dc99476706df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f07438e20bdf71c752bb661084c835341999c561bb5442d75e177223881276f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:58:51Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:58:43.431198 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:58:43.432227 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99450334/tls.crt::/tmp/serving-cert-99450334/tls.key\\\\\\\"\\\\nI0126 16:58:50.828320 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:58:50.830449 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:58:50.830472 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:58:50.830509 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:58:50.830515 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:58:50.834885 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:58:50.834958 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:58:50.834991 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0126 16:58:50.834905 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0126 16:58:50.835015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:58:50.835084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:58:50.835106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:58:50.835116 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0126 16:58:50.837641 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2daa3f13d37b7ae19dfca406b6b5cfdcd4f211287f7d410193f2fee36a24553\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:55Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.529112 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191ad5d41024c88c1a4fbc30f307dd6340fd55b93b00d22f62209d0e82be286f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2a4d8be409a46ffe5702797e56a80023a72a19fb7dc5c49e5f4984cedc600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:55Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.541478 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c823748751e9938f10fab08e33b5fffff5a6d15961ce028204131b7b69a56c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:55Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.543729 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 01:32:02.463317825 +0000 UTC Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.557384 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:55Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.565794 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.565827 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.565839 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.565855 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.565866 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:55Z","lastTransitionTime":"2026-01-26T16:59:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.573268 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad7b59f9-beb7-49d6-a2d1-e29133e46854\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53a6a6d4222ee5e1ac29fd5957d8cdf3fd42de72c68c85329374ce7afc4004\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v2l7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:55Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.593039 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71cb66e6f52823e7aa63f88ba1d153fde73816120aa75d4a6b910937303d2b9e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71cb66e6f52823e7aa63f88ba1d153fde73816120aa75d4a6b910937303d2b9e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:59:29Z\\\",\\\"message\\\":\\\"n-kubernetes/ovnkube-node-pxh94\\\\nI0126 16:59:29.834990 6550 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-t4fq2\\\\nI0126 16:59:29.834996 6550 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-t4fq2\\\\nI0126 16:59:29.835005 6550 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-t4fq2 in node crc\\\\nI0126 16:59:29.835026 6550 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0126 16:59:29.835056 6550 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0126 16:59:29.835079 6550 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0126 16:59:29.835090 6550 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nF0126 16:59:29.835097 6550 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pxh94_openshift-ovn-kubernetes(ab5b6f50-172b-4535-a0f9-5d103bcab4e7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pxh94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:55Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.604709 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-295wr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12e50462-28e6-4531-ada4-e652310e6cce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-295wr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:55Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.619590 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc864b0d-83bc-4954-9c61-ad650157caff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbde934a6c8acad10ca3ab8206d0ddbd4f7b17e9d304b898a68f4d3b0303bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b48f763d4aff37169399be766d5ab4f7ebbf91f304d139c9022a8556946eb107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0094e662f53c4832a984e05a880021af05ffc4c27f25394c28a070d9ef5490d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb95f623fa4f5f42217649ccce4225f9fe588bb3558da9334eef2b40ddb62486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb95f623fa4f5f42217649ccce4225f9fe588bb3558da9334eef2b40ddb62486\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:55Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.630890 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c3b0574-b4cc-483d-ae88-6517d1f30772\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec9063a7c03990fc26fc47427f164a769fd649c2bdbd9d23ea7f646e569734be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67797295a8c3952902b2696c6fdb26b72ce1826b5ccd522a24aac90a0411b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f67797295a8c3952902b2696c6fdb26b72ce1826b5ccd522a24aac90a0411b5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:55Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.642994 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63c75ede-5170-4db0-811b-5217ef8d72b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da26ddcfc6ccde3c9aabd63bdef3435f9a9eaab8c095bfeef5670c1295576cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54ca3fa13d9e8d442efa93b44a870369f7df3fe7562d77b98528f5c19a751f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xm9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:55Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.668541 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.668621 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.668636 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.668663 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.668679 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:55Z","lastTransitionTime":"2026-01-26T16:59:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.770914 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.770956 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.770972 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.770988 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.771001 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:55Z","lastTransitionTime":"2026-01-26T16:59:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.874080 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.874137 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.874148 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.874167 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.874200 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:55Z","lastTransitionTime":"2026-01-26T16:59:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.978816 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.978868 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.978883 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.978905 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:55 crc kubenswrapper[4856]: I0126 16:59:55.978921 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:55Z","lastTransitionTime":"2026-01-26T16:59:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.082661 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.082715 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.082732 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.082757 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.082774 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:56Z","lastTransitionTime":"2026-01-26T16:59:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.185373 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.185418 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.185428 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.185441 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.185450 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:56Z","lastTransitionTime":"2026-01-26T16:59:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.287936 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.288003 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.288026 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.288057 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.288078 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:56Z","lastTransitionTime":"2026-01-26T16:59:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.391286 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.391345 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.391356 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.391376 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.391390 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:56Z","lastTransitionTime":"2026-01-26T16:59:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.394645 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.394665 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.394723 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.394822 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 16:59:56 crc kubenswrapper[4856]: E0126 16:59:56.394901 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:59:56 crc kubenswrapper[4856]: E0126 16:59:56.395027 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:59:56 crc kubenswrapper[4856]: E0126 16:59:56.395104 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 16:59:56 crc kubenswrapper[4856]: E0126 16:59:56.395736 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.396101 4856 scope.go:117] "RemoveContainer" containerID="71cb66e6f52823e7aa63f88ba1d153fde73816120aa75d4a6b910937303d2b9e" Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.495098 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.495576 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.495594 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.495619 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.495641 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:56Z","lastTransitionTime":"2026-01-26T16:59:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.543882 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 21:28:58.629895815 +0000 UTC Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.598332 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.598365 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.598374 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.598388 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.598397 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:56Z","lastTransitionTime":"2026-01-26T16:59:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.700724 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.700797 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.700820 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.700848 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.700869 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:56Z","lastTransitionTime":"2026-01-26T16:59:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.803639 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.803709 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.803731 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.803757 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.803776 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:56Z","lastTransitionTime":"2026-01-26T16:59:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.907142 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.907224 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.907236 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.907258 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:56 crc kubenswrapper[4856]: I0126 16:59:56.907272 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:56Z","lastTransitionTime":"2026-01-26T16:59:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.011799 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.011890 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.011904 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.011929 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.011944 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:57Z","lastTransitionTime":"2026-01-26T16:59:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.114969 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.115045 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.115058 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.115077 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.115087 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:57Z","lastTransitionTime":"2026-01-26T16:59:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.218195 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.218256 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.218271 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.218290 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.218303 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:57Z","lastTransitionTime":"2026-01-26T16:59:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.321494 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.321580 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.321594 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.321617 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.321634 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:57Z","lastTransitionTime":"2026-01-26T16:59:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.424640 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.424702 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.424725 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.424750 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.424766 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:57Z","lastTransitionTime":"2026-01-26T16:59:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.527503 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.527568 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.527583 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.527603 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.527617 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:57Z","lastTransitionTime":"2026-01-26T16:59:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.544748 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 01:11:39.460528362 +0000 UTC Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.630805 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.630859 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.630875 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.630897 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.630915 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:57Z","lastTransitionTime":"2026-01-26T16:59:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.733810 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.733871 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.733884 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.733906 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.733921 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:57Z","lastTransitionTime":"2026-01-26T16:59:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.837068 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.837120 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.837128 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.837145 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.837157 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:57Z","lastTransitionTime":"2026-01-26T16:59:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.851916 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pxh94_ab5b6f50-172b-4535-a0f9-5d103bcab4e7/ovnkube-controller/2.log" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.854488 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" event={"ID":"ab5b6f50-172b-4535-a0f9-5d103bcab4e7","Type":"ContainerStarted","Data":"203d756903498bed1c57a3e87a95f4b24f808514567a99505ba2f3cfec468cb6"} Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.855477 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.869709 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af4930ff9cebd62106545fb3da2dfc93f7b591426fe4c85aa2e637b60c935f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:57Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.883513 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c09748c0963bdb50c9552f249d1135ea53e68c52babd788dad050951cb849cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:57Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.894950 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:57Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.906107 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v7579" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a77e85f9-b566-4807-bb92-55963c97b93c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba87c9fc35c230bbee201a5176cb467309f0b9aee82dfc81f3b677a15486d02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c03dc794e9c2035f2e1983eacad3e51d76223cb1b82e2f402c73f9453e4bd2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v7579\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:57Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.919113 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ecd87a-c5db-446d-ad3e-cfabbd648c1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89975b8f9428f81ab5d3fb48ced5dd9c837bea2feea3b89f5f7ff8d7d5d15b3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d432159a07ba20bf95f058bf8a597d67cbd8d852519bed035694f8ba3d8ec4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59ba0f131eee14df389391e46bc772894e49e976c3663df5fb3bc98b3cb4a3d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03a01b651d9f66da4dd1f6e0d29ad97c0d6ae46b644c3d997d8dc99476706df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f07438e20bdf71c752bb661084c835341999c561bb5442d75e177223881276f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:58:51Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:58:43.431198 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:58:43.432227 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99450334/tls.crt::/tmp/serving-cert-99450334/tls.key\\\\\\\"\\\\nI0126 16:58:50.828320 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:58:50.830449 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:58:50.830472 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:58:50.830509 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:58:50.830515 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:58:50.834885 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:58:50.834958 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:58:50.834991 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0126 16:58:50.834905 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0126 16:58:50.835015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:58:50.835084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:58:50.835106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:58:50.835116 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0126 16:58:50.837641 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2daa3f13d37b7ae19dfca406b6b5cfdcd4f211287f7d410193f2fee36a24553\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:57Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.934518 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191ad5d41024c88c1a4fbc30f307dd6340fd55b93b00d22f62209d0e82be286f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2a4d8be409a46ffe5702797e56a80023a72a19fb7dc5c49e5f4984cedc600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:57Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.939733 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.939790 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.939808 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.939832 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.939850 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:57Z","lastTransitionTime":"2026-01-26T16:59:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.949559 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c823748751e9938f10fab08e33b5fffff5a6d15961ce028204131b7b69a56c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:57Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.969257 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:57Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:57 crc kubenswrapper[4856]: I0126 16:59:57.987381 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://627cd6b39fdedd967fffcfc0755439277b2d73016fba1ddc7f5b9d5deba43b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:57Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.007566 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rq622" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a742e7b-c420-46e3-9e96-e9c744af6124\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afeb20035224feeab28a92ac77b43a24e653e49c56a25590a9861019a2b7a8ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad7222c9b91c0065a545bc1904d9864a5923cc13bfb6617daeb4a965a830f191\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:59:45Z\\\",\\\"message\\\":\\\"2026-01-26T16:59:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_41ab0694-d9c8-49a7-bf30-57e732ac7550\\\\n2026-01-26T16:59:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_41ab0694-d9c8-49a7-bf30-57e732ac7550 to /host/opt/cni/bin/\\\\n2026-01-26T16:59:00Z [verbose] multus-daemon started\\\\n2026-01-26T16:59:00Z [verbose] Readiness Indicator file check\\\\n2026-01-26T16:59:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8plh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rq622\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.027041 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.042929 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.042976 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.042987 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.043004 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.043015 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:58Z","lastTransitionTime":"2026-01-26T16:59:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.047074 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad7b59f9-beb7-49d6-a2d1-e29133e46854\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53a6a6d4222ee5e1ac29fd5957d8cdf3fd42de72c68c85329374ce7afc4004\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v2l7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.069498 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://203d756903498bed1c57a3e87a95f4b24f808514567a99505ba2f3cfec468cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71cb66e6f52823e7aa63f88ba1d153fde73816120aa75d4a6b910937303d2b9e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:59:29Z\\\",\\\"message\\\":\\\"n-kubernetes/ovnkube-node-pxh94\\\\nI0126 16:59:29.834990 6550 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-t4fq2\\\\nI0126 16:59:29.834996 6550 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-t4fq2\\\\nI0126 16:59:29.835005 6550 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-t4fq2 in node crc\\\\nI0126 16:59:29.835026 6550 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0126 16:59:29.835056 6550 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0126 16:59:29.835079 6550 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0126 16:59:29.835090 6550 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nF0126 16:59:29.835097 6550 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pxh94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.083121 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-295wr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12e50462-28e6-4531-ada4-e652310e6cce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-295wr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.099830 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc864b0d-83bc-4954-9c61-ad650157caff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbde934a6c8acad10ca3ab8206d0ddbd4f7b17e9d304b898a68f4d3b0303bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b48f763d4aff37169399be766d5ab4f7ebbf91f304d139c9022a8556946eb107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0094e662f53c4832a984e05a880021af05ffc4c27f25394c28a070d9ef5490d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb95f623fa4f5f42217649ccce4225f9fe588bb3558da9334eef2b40ddb62486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb95f623fa4f5f42217649ccce4225f9fe588bb3558da9334eef2b40ddb62486\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.111360 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c3b0574-b4cc-483d-ae88-6517d1f30772\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec9063a7c03990fc26fc47427f164a769fd649c2bdbd9d23ea7f646e569734be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67797295a8c3952902b2696c6fdb26b72ce1826b5ccd522a24aac90a0411b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f67797295a8c3952902b2696c6fdb26b72ce1826b5ccd522a24aac90a0411b5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.127104 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63c75ede-5170-4db0-811b-5217ef8d72b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da26ddcfc6ccde3c9aabd63bdef3435f9a9eaab8c095bfeef5670c1295576cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54ca3fa13d9e8d442efa93b44a870369f7df3fe7562d77b98528f5c19a751f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xm9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:58Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.145920 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.145953 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.145963 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.145981 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.146000 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:58Z","lastTransitionTime":"2026-01-26T16:59:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.248652 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.248711 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.248721 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.248739 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.248754 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:58Z","lastTransitionTime":"2026-01-26T16:59:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.351927 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.351979 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.351990 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.352011 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.352026 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:58Z","lastTransitionTime":"2026-01-26T16:59:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.394778 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.394819 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.394878 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:59:58 crc kubenswrapper[4856]: E0126 16:59:58.394939 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 16:59:58 crc kubenswrapper[4856]: E0126 16:59:58.395123 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.395174 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:59:58 crc kubenswrapper[4856]: E0126 16:59:58.395258 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 16:59:58 crc kubenswrapper[4856]: E0126 16:59:58.395320 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.455365 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.455445 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.455469 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.455498 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.455520 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:58Z","lastTransitionTime":"2026-01-26T16:59:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.545367 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 10:45:51.581499729 +0000 UTC Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.558907 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.558984 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.559009 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.559037 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.559058 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:58Z","lastTransitionTime":"2026-01-26T16:59:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.660980 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.661035 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.661051 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.661076 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.661096 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:58Z","lastTransitionTime":"2026-01-26T16:59:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.699629 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 16:59:58 crc kubenswrapper[4856]: E0126 16:59:58.699792 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:02.699760151 +0000 UTC m=+158.653014172 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.699845 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.699929 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.699992 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 16:59:58 crc kubenswrapper[4856]: E0126 16:59:58.700014 4856 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.700041 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 16:59:58 crc kubenswrapper[4856]: E0126 16:59:58.700064 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 17:01:02.70005169 +0000 UTC m=+158.653305671 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 16:59:58 crc kubenswrapper[4856]: E0126 16:59:58.700185 4856 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 16:59:58 crc kubenswrapper[4856]: E0126 16:59:58.700202 4856 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 16:59:58 crc kubenswrapper[4856]: E0126 16:59:58.700280 4856 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 16:59:58 crc kubenswrapper[4856]: E0126 16:59:58.700321 4856 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 16:59:58 crc kubenswrapper[4856]: E0126 16:59:58.700341 4856 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:59:58 crc kubenswrapper[4856]: E0126 16:59:58.700377 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 17:01:02.700336679 +0000 UTC m=+158.653590700 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 16:59:58 crc kubenswrapper[4856]: E0126 16:59:58.700409 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-26 17:01:02.700390281 +0000 UTC m=+158.653644302 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:59:58 crc kubenswrapper[4856]: E0126 16:59:58.700223 4856 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 16:59:58 crc kubenswrapper[4856]: E0126 16:59:58.700459 4856 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:59:58 crc kubenswrapper[4856]: E0126 16:59:58.700568 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-26 17:01:02.700519694 +0000 UTC m=+158.653773715 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.763452 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.763495 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.763506 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.763562 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.763576 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:58Z","lastTransitionTime":"2026-01-26T16:59:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.865867 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.865915 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.865944 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.865961 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.865975 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:58Z","lastTransitionTime":"2026-01-26T16:59:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.969160 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.969241 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.969264 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.969299 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:58 crc kubenswrapper[4856]: I0126 16:59:58.969326 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:58Z","lastTransitionTime":"2026-01-26T16:59:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.072786 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.072863 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.072875 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.072892 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.072904 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:59Z","lastTransitionTime":"2026-01-26T16:59:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.175853 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.175931 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.175948 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.175979 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.175998 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:59Z","lastTransitionTime":"2026-01-26T16:59:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.279270 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.279336 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.279357 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.279387 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.279411 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:59Z","lastTransitionTime":"2026-01-26T16:59:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.382575 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.382657 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.382675 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.382706 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.382725 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:59Z","lastTransitionTime":"2026-01-26T16:59:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.486600 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.486663 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.486681 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.486705 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.486722 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:59Z","lastTransitionTime":"2026-01-26T16:59:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.545972 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 01:17:42.634479283 +0000 UTC Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.590883 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.590918 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.590928 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.590944 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.590953 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:59Z","lastTransitionTime":"2026-01-26T16:59:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.694307 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.694365 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.694382 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.694405 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.694423 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:59Z","lastTransitionTime":"2026-01-26T16:59:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.797730 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.797757 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.797765 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.797779 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.797788 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:59Z","lastTransitionTime":"2026-01-26T16:59:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.865450 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pxh94_ab5b6f50-172b-4535-a0f9-5d103bcab4e7/ovnkube-controller/3.log" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.866648 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pxh94_ab5b6f50-172b-4535-a0f9-5d103bcab4e7/ovnkube-controller/2.log" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.871510 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.871640 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.871669 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.871706 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.871732 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:59Z","lastTransitionTime":"2026-01-26T16:59:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.873784 4856 generic.go:334] "Generic (PLEG): container finished" podID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerID="203d756903498bed1c57a3e87a95f4b24f808514567a99505ba2f3cfec468cb6" exitCode=1 Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.873862 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" event={"ID":"ab5b6f50-172b-4535-a0f9-5d103bcab4e7","Type":"ContainerDied","Data":"203d756903498bed1c57a3e87a95f4b24f808514567a99505ba2f3cfec468cb6"} Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.874231 4856 scope.go:117] "RemoveContainer" containerID="71cb66e6f52823e7aa63f88ba1d153fde73816120aa75d4a6b910937303d2b9e" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.875386 4856 scope.go:117] "RemoveContainer" containerID="203d756903498bed1c57a3e87a95f4b24f808514567a99505ba2f3cfec468cb6" Jan 26 16:59:59 crc kubenswrapper[4856]: E0126 16:59:59.875636 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pxh94_openshift-ovn-kubernetes(ab5b6f50-172b-4535-a0f9-5d103bcab4e7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" Jan 26 16:59:59 crc kubenswrapper[4856]: E0126 16:59:59.896262 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17523591-a778-4a97-aeab-8a7a93101850\\\",\\\"systemUUID\\\":\\\"ca45d056-99cb-4442-8a44-7e899628ecb2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:59Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.899825 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc864b0d-83bc-4954-9c61-ad650157caff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbde934a6c8acad10ca3ab8206d0ddbd4f7b17e9d304b898a68f4d3b0303bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b48f763d4aff37169399be766d5ab4f7ebbf91f304d139c9022a8556946eb107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0094e662f53c4832a984e05a880021af05ffc4c27f25394c28a070d9ef5490d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb95f623fa4f5f42217649ccce4225f9fe588bb3558da9334eef2b40ddb62486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb95f623fa4f5f42217649ccce4225f9fe588bb3558da9334eef2b40ddb62486\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:59Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.906110 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.906210 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.906228 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.906280 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.906295 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:59Z","lastTransitionTime":"2026-01-26T16:59:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.921460 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c3b0574-b4cc-483d-ae88-6517d1f30772\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec9063a7c03990fc26fc47427f164a769fd649c2bdbd9d23ea7f646e569734be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67797295a8c3952902b2696c6fdb26b72ce1826b5ccd522a24aac90a0411b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f67797295a8c3952902b2696c6fdb26b72ce1826b5ccd522a24aac90a0411b5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:59Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:59 crc kubenswrapper[4856]: E0126 16:59:59.928165 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17523591-a778-4a97-aeab-8a7a93101850\\\",\\\"systemUUID\\\":\\\"ca45d056-99cb-4442-8a44-7e899628ecb2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:59Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.934203 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.934275 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.934298 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.934326 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.934344 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:59Z","lastTransitionTime":"2026-01-26T16:59:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.939051 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63c75ede-5170-4db0-811b-5217ef8d72b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da26ddcfc6ccde3c9aabd63bdef3435f9a9eaab8c095bfeef5670c1295576cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54ca3fa13d9e8d442efa93b44a870369f7df3fe7562d77b98528f5c19a751f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xm9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:59Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.955359 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:59Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:59 crc kubenswrapper[4856]: E0126 16:59:59.956559 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17523591-a778-4a97-aeab-8a7a93101850\\\",\\\"systemUUID\\\":\\\"ca45d056-99cb-4442-8a44-7e899628ecb2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:59Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.961870 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.961909 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.961921 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.961937 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.961946 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:59Z","lastTransitionTime":"2026-01-26T16:59:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.971278 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v7579" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a77e85f9-b566-4807-bb92-55963c97b93c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba87c9fc35c230bbee201a5176cb467309f0b9aee82dfc81f3b677a15486d02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c03dc794e9c2035f2e1983eacad3e51d76223cb1b82e2f402c73f9453e4bd2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v7579\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:59Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:59 crc kubenswrapper[4856]: E0126 16:59:59.974816 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17523591-a778-4a97-aeab-8a7a93101850\\\",\\\"systemUUID\\\":\\\"ca45d056-99cb-4442-8a44-7e899628ecb2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:59Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.982095 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af4930ff9cebd62106545fb3da2dfc93f7b591426fe4c85aa2e637b60c935f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:59Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.982976 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.983005 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.983017 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.983034 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.983044 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T16:59:59Z","lastTransitionTime":"2026-01-26T16:59:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 16:59:59 crc kubenswrapper[4856]: I0126 16:59:59.997247 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c09748c0963bdb50c9552f249d1135ea53e68c52babd788dad050951cb849cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:59Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:59 crc kubenswrapper[4856]: E0126 16:59:59.998899 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T16:59:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17523591-a778-4a97-aeab-8a7a93101850\\\",\\\"systemUUID\\\":\\\"ca45d056-99cb-4442-8a44-7e899628ecb2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T16:59:59Z is after 2025-08-24T17:21:41Z" Jan 26 16:59:59 crc kubenswrapper[4856]: E0126 16:59:59.999001 4856 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.000859 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.000876 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.000884 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.000896 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.000905 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:00Z","lastTransitionTime":"2026-01-26T17:00:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.009929 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c823748751e9938f10fab08e33b5fffff5a6d15961ce028204131b7b69a56c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:00Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.026492 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:00Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.038414 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://627cd6b39fdedd967fffcfc0755439277b2d73016fba1ddc7f5b9d5deba43b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:00Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.074308 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rq622" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a742e7b-c420-46e3-9e96-e9c744af6124\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afeb20035224feeab28a92ac77b43a24e653e49c56a25590a9861019a2b7a8ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad7222c9b91c0065a545bc1904d9864a5923cc13bfb6617daeb4a965a830f191\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:59:45Z\\\",\\\"message\\\":\\\"2026-01-26T16:59:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_41ab0694-d9c8-49a7-bf30-57e732ac7550\\\\n2026-01-26T16:59:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_41ab0694-d9c8-49a7-bf30-57e732ac7550 to /host/opt/cni/bin/\\\\n2026-01-26T16:59:00Z [verbose] multus-daemon started\\\\n2026-01-26T16:59:00Z [verbose] Readiness Indicator file check\\\\n2026-01-26T16:59:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8plh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rq622\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:00Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.100613 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ecd87a-c5db-446d-ad3e-cfabbd648c1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89975b8f9428f81ab5d3fb48ced5dd9c837bea2feea3b89f5f7ff8d7d5d15b3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d432159a07ba20bf95f058bf8a597d67cbd8d852519bed035694f8ba3d8ec4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59ba0f131eee14df389391e46bc772894e49e976c3663df5fb3bc98b3cb4a3d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03a01b651d9f66da4dd1f6e0d29ad97c0d6ae46b644c3d997d8dc99476706df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f07438e20bdf71c752bb661084c835341999c561bb5442d75e177223881276f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:58:51Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:58:43.431198 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:58:43.432227 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99450334/tls.crt::/tmp/serving-cert-99450334/tls.key\\\\\\\"\\\\nI0126 16:58:50.828320 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:58:50.830449 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:58:50.830472 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:58:50.830509 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:58:50.830515 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:58:50.834885 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:58:50.834958 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:58:50.834991 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0126 16:58:50.834905 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0126 16:58:50.835015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:58:50.835084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:58:50.835106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:58:50.835116 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0126 16:58:50.837641 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2daa3f13d37b7ae19dfca406b6b5cfdcd4f211287f7d410193f2fee36a24553\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:00Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.103592 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.103637 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.103649 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.103669 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.103681 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:00Z","lastTransitionTime":"2026-01-26T17:00:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.122286 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191ad5d41024c88c1a4fbc30f307dd6340fd55b93b00d22f62209d0e82be286f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2a4d8be409a46ffe5702797e56a80023a72a19fb7dc5c49e5f4984cedc600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:00Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.138237 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-295wr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12e50462-28e6-4531-ada4-e652310e6cce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-295wr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:00Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.159734 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:00Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.177362 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad7b59f9-beb7-49d6-a2d1-e29133e46854\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53a6a6d4222ee5e1ac29fd5957d8cdf3fd42de72c68c85329374ce7afc4004\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v2l7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:00Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.198750 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://203d756903498bed1c57a3e87a95f4b24f808514567a99505ba2f3cfec468cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71cb66e6f52823e7aa63f88ba1d153fde73816120aa75d4a6b910937303d2b9e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:59:29Z\\\",\\\"message\\\":\\\"n-kubernetes/ovnkube-node-pxh94\\\\nI0126 16:59:29.834990 6550 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-t4fq2\\\\nI0126 16:59:29.834996 6550 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-t4fq2\\\\nI0126 16:59:29.835005 6550 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-t4fq2 in node crc\\\\nI0126 16:59:29.835026 6550 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0126 16:59:29.835056 6550 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0126 16:59:29.835079 6550 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0126 16:59:29.835090 6550 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nF0126 16:59:29.835097 6550 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://203d756903498bed1c57a3e87a95f4b24f808514567a99505ba2f3cfec468cb6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:59:58Z\\\",\\\"message\\\":\\\" server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0126 16:59:57.729092 6886 port_cache.go:96] port-cache(openshift-network-diagnostics_network-check-target-xd92c): added port \\\\u0026{name:openshift-network-diagnostics_network-check-target-xd92c uuid:61897e97-c771-4738-8709-09636387cb00 logicalSwitch:crc ips:[0xc008f54210] mac:[10 88 10 217 0 4] expires:{wall:0 ext:0 loc:\\\\u003cnil\\\\u003e}} with IP: [10.217.0.4/23] and MAC: 0a:58:0a:d9:00:04\\\\nI0126 16:59:57.729124 6886 pods.go:252] [openshift-network-diagnostics/network-check-target-xd92c] addLogicalPort took 1.706822ms, libovsdb time 885.597µs\\\\nI0126 16:59:57.729132 6886 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-diagnostics/network-check-target-xd92c after 0 failed attempt(s)\\\\nF0126 16:59:57.729134 6886 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": fa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pxh94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:00Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.206429 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.206494 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.206512 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.206582 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.206600 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:00Z","lastTransitionTime":"2026-01-26T17:00:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.308956 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.308985 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.308995 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.309010 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.309020 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:00Z","lastTransitionTime":"2026-01-26T17:00:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.395005 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.395066 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.395105 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.395080 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 17:00:00 crc kubenswrapper[4856]: E0126 17:00:00.395150 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 17:00:00 crc kubenswrapper[4856]: E0126 17:00:00.395222 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 17:00:00 crc kubenswrapper[4856]: E0126 17:00:00.395273 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 17:00:00 crc kubenswrapper[4856]: E0126 17:00:00.395325 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.411122 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.411167 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.411180 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.411196 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.411209 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:00Z","lastTransitionTime":"2026-01-26T17:00:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.514740 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.514804 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.514828 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.514881 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.514906 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:00Z","lastTransitionTime":"2026-01-26T17:00:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.546266 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 09:12:06.336930725 +0000 UTC Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.621669 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.621792 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.621812 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.622437 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.622482 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:00Z","lastTransitionTime":"2026-01-26T17:00:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.725215 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.725255 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.725265 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.725278 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.725288 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:00Z","lastTransitionTime":"2026-01-26T17:00:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.828247 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.828297 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.828310 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.828328 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.828339 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:00Z","lastTransitionTime":"2026-01-26T17:00:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.880783 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pxh94_ab5b6f50-172b-4535-a0f9-5d103bcab4e7/ovnkube-controller/3.log" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.930928 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.931000 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.931020 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.931047 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:00 crc kubenswrapper[4856]: I0126 17:00:00.931065 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:00Z","lastTransitionTime":"2026-01-26T17:00:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.033924 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.033965 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.033975 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.034017 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.034030 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:01Z","lastTransitionTime":"2026-01-26T17:00:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.136194 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.136235 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.136246 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.136263 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.136280 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:01Z","lastTransitionTime":"2026-01-26T17:00:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.238849 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.238884 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.238893 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.238906 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.238915 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:01Z","lastTransitionTime":"2026-01-26T17:00:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.341090 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.341476 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.341492 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.341510 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.341520 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:01Z","lastTransitionTime":"2026-01-26T17:00:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.444217 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.444259 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.444272 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.444289 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.444300 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:01Z","lastTransitionTime":"2026-01-26T17:00:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.546061 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.546098 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.546109 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.546124 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.546135 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:01Z","lastTransitionTime":"2026-01-26T17:00:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.546453 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 08:14:26.676118664 +0000 UTC Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.648585 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.648626 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.648635 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.648650 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.648660 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:01Z","lastTransitionTime":"2026-01-26T17:00:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.752058 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.752108 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.752122 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.752144 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.752159 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:01Z","lastTransitionTime":"2026-01-26T17:00:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.854222 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.854263 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.854272 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.854286 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.854295 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:01Z","lastTransitionTime":"2026-01-26T17:00:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.958120 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.958166 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.958181 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.958202 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:01 crc kubenswrapper[4856]: I0126 17:00:01.958217 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:01Z","lastTransitionTime":"2026-01-26T17:00:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.060832 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.060874 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.060883 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.060898 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.060908 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:02Z","lastTransitionTime":"2026-01-26T17:00:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.163403 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.163468 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.163483 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.163504 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.163551 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:02Z","lastTransitionTime":"2026-01-26T17:00:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.266205 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.266261 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.266272 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.266292 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.266309 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:02Z","lastTransitionTime":"2026-01-26T17:00:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.369071 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.369134 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.369152 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.369175 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.369192 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:02Z","lastTransitionTime":"2026-01-26T17:00:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.394351 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.394444 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.394477 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.394514 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 17:00:02 crc kubenswrapper[4856]: E0126 17:00:02.394709 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 17:00:02 crc kubenswrapper[4856]: E0126 17:00:02.394868 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 17:00:02 crc kubenswrapper[4856]: E0126 17:00:02.394977 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 17:00:02 crc kubenswrapper[4856]: E0126 17:00:02.395130 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.471883 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.471932 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.471949 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.471968 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.471983 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:02Z","lastTransitionTime":"2026-01-26T17:00:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.547602 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 05:07:47.139019889 +0000 UTC Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.574797 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.574839 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.574849 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.574867 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.574879 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:02Z","lastTransitionTime":"2026-01-26T17:00:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.678332 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.678395 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.678416 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.678442 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.678458 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:02Z","lastTransitionTime":"2026-01-26T17:00:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.781743 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.781807 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.781825 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.781851 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.781869 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:02Z","lastTransitionTime":"2026-01-26T17:00:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.885206 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.885278 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.885312 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.885341 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.885362 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:02Z","lastTransitionTime":"2026-01-26T17:00:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.988256 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.988312 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.988330 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.988356 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:02 crc kubenswrapper[4856]: I0126 17:00:02.988372 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:02Z","lastTransitionTime":"2026-01-26T17:00:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.090379 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.090420 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.090432 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.090448 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.090459 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:03Z","lastTransitionTime":"2026-01-26T17:00:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.192964 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.193012 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.193023 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.193041 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.193053 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:03Z","lastTransitionTime":"2026-01-26T17:00:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.296316 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.296366 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.296377 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.296393 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.296403 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:03Z","lastTransitionTime":"2026-01-26T17:00:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.399778 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.399815 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.399827 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.399855 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.399868 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:03Z","lastTransitionTime":"2026-01-26T17:00:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.502831 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.502868 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.502886 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.502905 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.502916 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:03Z","lastTransitionTime":"2026-01-26T17:00:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.548111 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 17:49:15.869698335 +0000 UTC Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.605949 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.606005 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.606017 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.606034 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.606046 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:03Z","lastTransitionTime":"2026-01-26T17:00:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.709396 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.709480 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.709506 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.709580 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.709620 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:03Z","lastTransitionTime":"2026-01-26T17:00:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.813184 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.813222 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.813234 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.813250 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.813262 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:03Z","lastTransitionTime":"2026-01-26T17:00:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.915171 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.915272 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.915297 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.915327 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:03 crc kubenswrapper[4856]: I0126 17:00:03.915352 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:03Z","lastTransitionTime":"2026-01-26T17:00:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.018408 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.018462 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.018479 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.018500 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.018513 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:04Z","lastTransitionTime":"2026-01-26T17:00:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.121640 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.121713 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.121735 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.121763 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.121783 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:04Z","lastTransitionTime":"2026-01-26T17:00:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.224777 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.224828 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.224846 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.224868 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.224883 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:04Z","lastTransitionTime":"2026-01-26T17:00:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.328897 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.328971 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.328995 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.329025 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.329051 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:04Z","lastTransitionTime":"2026-01-26T17:00:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.394861 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.394898 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.394988 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 17:00:04 crc kubenswrapper[4856]: E0126 17:00:04.395076 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.395146 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 17:00:04 crc kubenswrapper[4856]: E0126 17:00:04.395247 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 17:00:04 crc kubenswrapper[4856]: E0126 17:00:04.395356 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 17:00:04 crc kubenswrapper[4856]: E0126 17:00:04.395516 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.431839 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.431896 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.431914 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.431936 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.431954 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:04Z","lastTransitionTime":"2026-01-26T17:00:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.534596 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.534655 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.534672 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.534693 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.534711 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:04Z","lastTransitionTime":"2026-01-26T17:00:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.549178 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 13:13:36.824011002 +0000 UTC Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.637481 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.637564 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.637581 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.637604 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.637620 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:04Z","lastTransitionTime":"2026-01-26T17:00:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.739749 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.739814 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.739831 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.739859 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.739877 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:04Z","lastTransitionTime":"2026-01-26T17:00:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.843139 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.843199 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.843238 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.843273 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.843291 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:04Z","lastTransitionTime":"2026-01-26T17:00:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.945776 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.945821 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.945831 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.945845 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:04 crc kubenswrapper[4856]: I0126 17:00:04.945853 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:04Z","lastTransitionTime":"2026-01-26T17:00:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.048555 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.048608 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.048624 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.048644 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.048658 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:05Z","lastTransitionTime":"2026-01-26T17:00:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.151305 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.151357 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.151369 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.151387 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.151400 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:05Z","lastTransitionTime":"2026-01-26T17:00:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.253505 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.253582 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.253599 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.253621 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.253638 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:05Z","lastTransitionTime":"2026-01-26T17:00:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.357125 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.357186 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.357206 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.357229 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.357245 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:05Z","lastTransitionTime":"2026-01-26T17:00:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.425619 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://203d756903498bed1c57a3e87a95f4b24f808514567a99505ba2f3cfec468cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71cb66e6f52823e7aa63f88ba1d153fde73816120aa75d4a6b910937303d2b9e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:59:29Z\\\",\\\"message\\\":\\\"n-kubernetes/ovnkube-node-pxh94\\\\nI0126 16:59:29.834990 6550 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-t4fq2\\\\nI0126 16:59:29.834996 6550 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-t4fq2\\\\nI0126 16:59:29.835005 6550 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-t4fq2 in node crc\\\\nI0126 16:59:29.835026 6550 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0126 16:59:29.835056 6550 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0126 16:59:29.835079 6550 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0126 16:59:29.835090 6550 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nF0126 16:59:29.835097 6550 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://203d756903498bed1c57a3e87a95f4b24f808514567a99505ba2f3cfec468cb6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:59:58Z\\\",\\\"message\\\":\\\" server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0126 16:59:57.729092 6886 port_cache.go:96] port-cache(openshift-network-diagnostics_network-check-target-xd92c): added port \\\\u0026{name:openshift-network-diagnostics_network-check-target-xd92c uuid:61897e97-c771-4738-8709-09636387cb00 logicalSwitch:crc ips:[0xc008f54210] mac:[10 88 10 217 0 4] expires:{wall:0 ext:0 loc:\\\\u003cnil\\\\u003e}} with IP: [10.217.0.4/23] and MAC: 0a:58:0a:d9:00:04\\\\nI0126 16:59:57.729124 6886 pods.go:252] [openshift-network-diagnostics/network-check-target-xd92c] addLogicalPort took 1.706822ms, libovsdb time 885.597µs\\\\nI0126 16:59:57.729132 6886 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-diagnostics/network-check-target-xd92c after 0 failed attempt(s)\\\\nF0126 16:59:57.729134 6886 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": fa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pxh94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:05Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.440410 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-295wr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12e50462-28e6-4531-ada4-e652310e6cce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-295wr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:05Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.455002 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:05Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.459580 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.459627 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.459641 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.459658 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.459669 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:05Z","lastTransitionTime":"2026-01-26T17:00:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.470359 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad7b59f9-beb7-49d6-a2d1-e29133e46854\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53a6a6d4222ee5e1ac29fd5957d8cdf3fd42de72c68c85329374ce7afc4004\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v2l7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:05Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.482355 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63c75ede-5170-4db0-811b-5217ef8d72b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da26ddcfc6ccde3c9aabd63bdef3435f9a9eaab8c095bfeef5670c1295576cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54ca3fa13d9e8d442efa93b44a870369f7df3fe7562d77b98528f5c19a751f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xm9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:05Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.494738 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc864b0d-83bc-4954-9c61-ad650157caff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbde934a6c8acad10ca3ab8206d0ddbd4f7b17e9d304b898a68f4d3b0303bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b48f763d4aff37169399be766d5ab4f7ebbf91f304d139c9022a8556946eb107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0094e662f53c4832a984e05a880021af05ffc4c27f25394c28a070d9ef5490d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb95f623fa4f5f42217649ccce4225f9fe588bb3558da9334eef2b40ddb62486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb95f623fa4f5f42217649ccce4225f9fe588bb3558da9334eef2b40ddb62486\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:05Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.505308 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c3b0574-b4cc-483d-ae88-6517d1f30772\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec9063a7c03990fc26fc47427f164a769fd649c2bdbd9d23ea7f646e569734be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67797295a8c3952902b2696c6fdb26b72ce1826b5ccd522a24aac90a0411b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f67797295a8c3952902b2696c6fdb26b72ce1826b5ccd522a24aac90a0411b5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:05Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.523200 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c09748c0963bdb50c9552f249d1135ea53e68c52babd788dad050951cb849cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:05Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.536482 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:05Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.548266 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v7579" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a77e85f9-b566-4807-bb92-55963c97b93c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba87c9fc35c230bbee201a5176cb467309f0b9aee82dfc81f3b677a15486d02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c03dc794e9c2035f2e1983eacad3e51d76223cb1b82e2f402c73f9453e4bd2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v7579\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:05Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.550359 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 01:54:03.416173542 +0000 UTC Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.559599 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af4930ff9cebd62106545fb3da2dfc93f7b591426fe4c85aa2e637b60c935f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:05Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.562228 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.562254 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.562262 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.562275 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.562284 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:05Z","lastTransitionTime":"2026-01-26T17:00:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.577692 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191ad5d41024c88c1a4fbc30f307dd6340fd55b93b00d22f62209d0e82be286f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2a4d8be409a46ffe5702797e56a80023a72a19fb7dc5c49e5f4984cedc600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:05Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.590702 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c823748751e9938f10fab08e33b5fffff5a6d15961ce028204131b7b69a56c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:05Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.606974 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:05Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.621714 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://627cd6b39fdedd967fffcfc0755439277b2d73016fba1ddc7f5b9d5deba43b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:05Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.638831 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rq622" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a742e7b-c420-46e3-9e96-e9c744af6124\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afeb20035224feeab28a92ac77b43a24e653e49c56a25590a9861019a2b7a8ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad7222c9b91c0065a545bc1904d9864a5923cc13bfb6617daeb4a965a830f191\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:59:45Z\\\",\\\"message\\\":\\\"2026-01-26T16:59:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_41ab0694-d9c8-49a7-bf30-57e732ac7550\\\\n2026-01-26T16:59:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_41ab0694-d9c8-49a7-bf30-57e732ac7550 to /host/opt/cni/bin/\\\\n2026-01-26T16:59:00Z [verbose] multus-daemon started\\\\n2026-01-26T16:59:00Z [verbose] Readiness Indicator file check\\\\n2026-01-26T16:59:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8plh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rq622\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:05Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.656304 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ecd87a-c5db-446d-ad3e-cfabbd648c1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89975b8f9428f81ab5d3fb48ced5dd9c837bea2feea3b89f5f7ff8d7d5d15b3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d432159a07ba20bf95f058bf8a597d67cbd8d852519bed035694f8ba3d8ec4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59ba0f131eee14df389391e46bc772894e49e976c3663df5fb3bc98b3cb4a3d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03a01b651d9f66da4dd1f6e0d29ad97c0d6ae46b644c3d997d8dc99476706df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f07438e20bdf71c752bb661084c835341999c561bb5442d75e177223881276f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:58:51Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:58:43.431198 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:58:43.432227 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99450334/tls.crt::/tmp/serving-cert-99450334/tls.key\\\\\\\"\\\\nI0126 16:58:50.828320 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:58:50.830449 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:58:50.830472 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:58:50.830509 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:58:50.830515 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:58:50.834885 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:58:50.834958 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:58:50.834991 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0126 16:58:50.834905 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0126 16:58:50.835015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:58:50.835084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:58:50.835106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:58:50.835116 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0126 16:58:50.837641 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2daa3f13d37b7ae19dfca406b6b5cfdcd4f211287f7d410193f2fee36a24553\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:05Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.664885 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.664937 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.664955 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.664978 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.664996 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:05Z","lastTransitionTime":"2026-01-26T17:00:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.767397 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.767466 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.767502 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.767519 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.767556 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:05Z","lastTransitionTime":"2026-01-26T17:00:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.871899 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.871966 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.871984 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.872013 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.872031 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:05Z","lastTransitionTime":"2026-01-26T17:00:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.976285 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.976322 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.976333 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.976348 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:05 crc kubenswrapper[4856]: I0126 17:00:05.976360 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:05Z","lastTransitionTime":"2026-01-26T17:00:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.079353 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.079415 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.079431 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.079451 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.079466 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:06Z","lastTransitionTime":"2026-01-26T17:00:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.182128 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.182204 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.182228 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.182260 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.182287 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:06Z","lastTransitionTime":"2026-01-26T17:00:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.285803 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.285867 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.285906 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.285940 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.285963 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:06Z","lastTransitionTime":"2026-01-26T17:00:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.388814 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.388887 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.388907 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.388931 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.388950 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:06Z","lastTransitionTime":"2026-01-26T17:00:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.395151 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.395185 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.395215 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 17:00:06 crc kubenswrapper[4856]: E0126 17:00:06.395283 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.395298 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 17:00:06 crc kubenswrapper[4856]: E0126 17:00:06.395464 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 17:00:06 crc kubenswrapper[4856]: E0126 17:00:06.395657 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 17:00:06 crc kubenswrapper[4856]: E0126 17:00:06.395796 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.491658 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.491858 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.491890 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.491914 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.491931 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:06Z","lastTransitionTime":"2026-01-26T17:00:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.551001 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 16:40:08.933451497 +0000 UTC Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.595163 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.595237 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.595256 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.595280 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.595300 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:06Z","lastTransitionTime":"2026-01-26T17:00:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.698025 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.698086 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.698097 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.698118 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.698129 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:06Z","lastTransitionTime":"2026-01-26T17:00:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.801059 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.801092 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.801100 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.801112 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.801121 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:06Z","lastTransitionTime":"2026-01-26T17:00:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.904168 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.904222 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.904239 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.904265 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:06 crc kubenswrapper[4856]: I0126 17:00:06.904284 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:06Z","lastTransitionTime":"2026-01-26T17:00:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.007476 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.007580 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.007608 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.007642 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.007666 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:07Z","lastTransitionTime":"2026-01-26T17:00:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.110492 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.110579 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.110615 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.110652 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.110673 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:07Z","lastTransitionTime":"2026-01-26T17:00:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.214125 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.214196 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.214221 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.214250 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.214272 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:07Z","lastTransitionTime":"2026-01-26T17:00:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.317334 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.317509 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.317571 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.317640 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.317666 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:07Z","lastTransitionTime":"2026-01-26T17:00:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.421797 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.422898 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.422958 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.422975 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.422999 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.423019 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:07Z","lastTransitionTime":"2026-01-26T17:00:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.526317 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.526367 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.526406 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.526438 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.526463 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:07Z","lastTransitionTime":"2026-01-26T17:00:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.551854 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 05:32:43.980500834 +0000 UTC Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.629297 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.629329 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.629337 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.629350 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.629358 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:07Z","lastTransitionTime":"2026-01-26T17:00:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.732011 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.732059 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.732069 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.732084 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.732095 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:07Z","lastTransitionTime":"2026-01-26T17:00:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.835096 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.835203 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.835238 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.835279 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.835303 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:07Z","lastTransitionTime":"2026-01-26T17:00:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.937599 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.937721 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.937737 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.937760 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:07 crc kubenswrapper[4856]: I0126 17:00:07.937776 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:07Z","lastTransitionTime":"2026-01-26T17:00:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.040092 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.040146 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.040164 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.040186 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.040202 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:08Z","lastTransitionTime":"2026-01-26T17:00:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.143513 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.143588 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.143600 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.143617 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.143630 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:08Z","lastTransitionTime":"2026-01-26T17:00:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.245221 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.245254 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.245264 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.245277 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.245286 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:08Z","lastTransitionTime":"2026-01-26T17:00:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.348825 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.348885 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.348904 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.348947 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.348973 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:08Z","lastTransitionTime":"2026-01-26T17:00:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.395465 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 17:00:08 crc kubenswrapper[4856]: E0126 17:00:08.406341 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.407129 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 17:00:08 crc kubenswrapper[4856]: E0126 17:00:08.407231 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.407275 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 17:00:08 crc kubenswrapper[4856]: E0126 17:00:08.407324 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.407359 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 17:00:08 crc kubenswrapper[4856]: E0126 17:00:08.407399 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.423717 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.451775 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.451814 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.451823 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.451843 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.451854 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:08Z","lastTransitionTime":"2026-01-26T17:00:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.552334 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 05:09:34.646750944 +0000 UTC Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.554351 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.554406 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.554422 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.554444 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.554459 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:08Z","lastTransitionTime":"2026-01-26T17:00:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.657566 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.657625 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.657640 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.657664 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.657678 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:08Z","lastTransitionTime":"2026-01-26T17:00:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.760865 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.760909 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.760919 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.760934 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.760944 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:08Z","lastTransitionTime":"2026-01-26T17:00:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.863951 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.864006 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.864024 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.864048 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.864060 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:08Z","lastTransitionTime":"2026-01-26T17:00:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.967177 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.967227 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.967237 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.967256 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:08 crc kubenswrapper[4856]: I0126 17:00:08.967267 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:08Z","lastTransitionTime":"2026-01-26T17:00:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.069346 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.069412 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.069424 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.069438 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.069447 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:09Z","lastTransitionTime":"2026-01-26T17:00:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.173030 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.173076 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.173088 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.173112 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.173126 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:09Z","lastTransitionTime":"2026-01-26T17:00:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.276496 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.276586 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.276602 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.276628 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.276649 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:09Z","lastTransitionTime":"2026-01-26T17:00:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.379572 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.379617 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.379632 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.379654 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.379672 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:09Z","lastTransitionTime":"2026-01-26T17:00:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.482739 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.482809 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.482818 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.482864 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.482877 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:09Z","lastTransitionTime":"2026-01-26T17:00:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.553034 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 13:46:06.645148637 +0000 UTC Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.586175 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.586248 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.586274 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.586304 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.586327 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:09Z","lastTransitionTime":"2026-01-26T17:00:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.691081 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.691183 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.691210 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.691260 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.691287 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:09Z","lastTransitionTime":"2026-01-26T17:00:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.794307 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.794351 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.794362 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.794375 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.794384 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:09Z","lastTransitionTime":"2026-01-26T17:00:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.897196 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.897307 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.897328 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.897351 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:09 crc kubenswrapper[4856]: I0126 17:00:09.897365 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:09Z","lastTransitionTime":"2026-01-26T17:00:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.000769 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.000820 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.000832 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.000855 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.000868 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:10Z","lastTransitionTime":"2026-01-26T17:00:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.104083 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.104150 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.104160 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.104182 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.104194 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:10Z","lastTransitionTime":"2026-01-26T17:00:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.115789 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.115860 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.115872 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.115894 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.115909 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:10Z","lastTransitionTime":"2026-01-26T17:00:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:10 crc kubenswrapper[4856]: E0126 17:00:10.131350 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T17:00:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T17:00:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T17:00:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T17:00:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T17:00:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T17:00:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T17:00:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T17:00:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17523591-a778-4a97-aeab-8a7a93101850\\\",\\\"systemUUID\\\":\\\"ca45d056-99cb-4442-8a44-7e899628ecb2\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:10Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.136316 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.136384 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.136411 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.136444 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.136473 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:10Z","lastTransitionTime":"2026-01-26T17:00:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:10 crc kubenswrapper[4856]: E0126 17:00:10.155757 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T17:00:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T17:00:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T17:00:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T17:00:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T17:00:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T17:00:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T17:00:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T17:00:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17523591-a778-4a97-aeab-8a7a93101850\\\",\\\"systemUUID\\\":\\\"ca45d056-99cb-4442-8a44-7e899628ecb2\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:10Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.161722 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.161755 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.161765 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.161781 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.161792 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:10Z","lastTransitionTime":"2026-01-26T17:00:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:10 crc kubenswrapper[4856]: E0126 17:00:10.177744 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T17:00:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T17:00:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T17:00:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T17:00:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T17:00:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T17:00:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T17:00:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T17:00:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17523591-a778-4a97-aeab-8a7a93101850\\\",\\\"systemUUID\\\":\\\"ca45d056-99cb-4442-8a44-7e899628ecb2\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:10Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.182417 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.182463 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.182475 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.182495 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.182508 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:10Z","lastTransitionTime":"2026-01-26T17:00:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:10 crc kubenswrapper[4856]: E0126 17:00:10.197713 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T17:00:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T17:00:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T17:00:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T17:00:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T17:00:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T17:00:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T17:00:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T17:00:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17523591-a778-4a97-aeab-8a7a93101850\\\",\\\"systemUUID\\\":\\\"ca45d056-99cb-4442-8a44-7e899628ecb2\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:10Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.202159 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.202199 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.202210 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.202228 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.202241 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:10Z","lastTransitionTime":"2026-01-26T17:00:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:10 crc kubenswrapper[4856]: E0126 17:00:10.216614 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T17:00:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T17:00:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T17:00:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T17:00:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T17:00:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T17:00:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T17:00:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T17:00:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17523591-a778-4a97-aeab-8a7a93101850\\\",\\\"systemUUID\\\":\\\"ca45d056-99cb-4442-8a44-7e899628ecb2\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:10Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:10 crc kubenswrapper[4856]: E0126 17:00:10.216765 4856 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.218254 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.218291 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.218302 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.218321 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.218332 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:10Z","lastTransitionTime":"2026-01-26T17:00:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.321684 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.321739 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.321752 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.321772 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.321790 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:10Z","lastTransitionTime":"2026-01-26T17:00:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.394803 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.394984 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.395005 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 17:00:10 crc kubenswrapper[4856]: E0126 17:00:10.395126 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 17:00:10 crc kubenswrapper[4856]: E0126 17:00:10.395302 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.395331 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 17:00:10 crc kubenswrapper[4856]: E0126 17:00:10.395508 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 17:00:10 crc kubenswrapper[4856]: E0126 17:00:10.395627 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.424279 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.424319 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.424330 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.424348 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.424360 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:10Z","lastTransitionTime":"2026-01-26T17:00:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.528630 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.528709 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.528729 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.528755 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.528786 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:10Z","lastTransitionTime":"2026-01-26T17:00:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.553630 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 02:57:46.617172638 +0000 UTC Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.631860 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.631902 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.631917 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.631942 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.631957 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:10Z","lastTransitionTime":"2026-01-26T17:00:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.734453 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.734560 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.734583 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.734800 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.734816 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:10Z","lastTransitionTime":"2026-01-26T17:00:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.837874 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.837938 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.837962 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.837994 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.838017 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:10Z","lastTransitionTime":"2026-01-26T17:00:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.941049 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.941125 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.941143 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.941166 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:10 crc kubenswrapper[4856]: I0126 17:00:10.941184 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:10Z","lastTransitionTime":"2026-01-26T17:00:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.044674 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.044755 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.044824 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.044849 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.044871 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:11Z","lastTransitionTime":"2026-01-26T17:00:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.146895 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.146942 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.146955 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.146972 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.146984 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:11Z","lastTransitionTime":"2026-01-26T17:00:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.249984 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.250053 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.250069 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.250095 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.250113 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:11Z","lastTransitionTime":"2026-01-26T17:00:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.353688 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.353750 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.353768 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.353791 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.353809 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:11Z","lastTransitionTime":"2026-01-26T17:00:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.456375 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.456457 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.456480 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.456689 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.456710 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:11Z","lastTransitionTime":"2026-01-26T17:00:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.554296 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 22:45:43.688075778 +0000 UTC Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.560263 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.560317 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.560336 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.560361 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.560436 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:11Z","lastTransitionTime":"2026-01-26T17:00:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.671203 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.671291 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.671317 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.671444 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.671473 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:11Z","lastTransitionTime":"2026-01-26T17:00:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.775059 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.775115 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.775127 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.775146 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.775162 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:11Z","lastTransitionTime":"2026-01-26T17:00:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.877924 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.877987 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.878004 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.878028 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.878047 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:11Z","lastTransitionTime":"2026-01-26T17:00:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.981951 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.982016 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.982033 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.982050 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:11 crc kubenswrapper[4856]: I0126 17:00:11.982060 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:11Z","lastTransitionTime":"2026-01-26T17:00:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.085778 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.085850 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.085864 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.085879 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.085889 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:12Z","lastTransitionTime":"2026-01-26T17:00:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.189218 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.189291 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.189309 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.189336 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.189354 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:12Z","lastTransitionTime":"2026-01-26T17:00:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.292226 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.292266 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.292274 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.292297 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.292307 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:12Z","lastTransitionTime":"2026-01-26T17:00:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.394441 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.394512 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.394574 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 17:00:12 crc kubenswrapper[4856]: E0126 17:00:12.394695 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 17:00:12 crc kubenswrapper[4856]: E0126 17:00:12.394891 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.394972 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 17:00:12 crc kubenswrapper[4856]: E0126 17:00:12.395032 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 17:00:12 crc kubenswrapper[4856]: E0126 17:00:12.395121 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.395727 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.395794 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.395815 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.395837 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.395854 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:12Z","lastTransitionTime":"2026-01-26T17:00:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.499392 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.499450 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.499471 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.499497 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.499514 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:12Z","lastTransitionTime":"2026-01-26T17:00:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.555217 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 18:01:13.908136797 +0000 UTC Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.602102 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.602185 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.602213 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.602248 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.602274 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:12Z","lastTransitionTime":"2026-01-26T17:00:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.706014 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.706075 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.706085 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.706105 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.706118 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:12Z","lastTransitionTime":"2026-01-26T17:00:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.811940 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.812037 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.812050 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.812087 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.812115 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:12Z","lastTransitionTime":"2026-01-26T17:00:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.915004 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.915054 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.915071 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.915090 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:12 crc kubenswrapper[4856]: I0126 17:00:12.915105 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:12Z","lastTransitionTime":"2026-01-26T17:00:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.018338 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.018394 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.018404 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.018423 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.018469 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:13Z","lastTransitionTime":"2026-01-26T17:00:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.121866 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.121935 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.121953 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.121980 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.121997 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:13Z","lastTransitionTime":"2026-01-26T17:00:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.225234 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.225288 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.225301 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.225371 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.225387 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:13Z","lastTransitionTime":"2026-01-26T17:00:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.328219 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.328312 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.328327 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.328347 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.328359 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:13Z","lastTransitionTime":"2026-01-26T17:00:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.397143 4856 scope.go:117] "RemoveContainer" containerID="203d756903498bed1c57a3e87a95f4b24f808514567a99505ba2f3cfec468cb6" Jan 26 17:00:13 crc kubenswrapper[4856]: E0126 17:00:13.397668 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pxh94_openshift-ovn-kubernetes(ab5b6f50-172b-4535-a0f9-5d103bcab4e7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.422757 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://191ad5d41024c88c1a4fbc30f307dd6340fd55b93b00d22f62209d0e82be286f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccf2a4d8be409a46ffe5702797e56a80023a72a19fb7dc5c49e5f4984cedc600\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:13Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.431391 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.431472 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.431496 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.431560 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.431586 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:13Z","lastTransitionTime":"2026-01-26T17:00:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.445153 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c823748751e9938f10fab08e33b5fffff5a6d15961ce028204131b7b69a56c18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:13Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.468896 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:13Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.488003 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-t4fq2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8d21ac89-2ebd-49c3-9fe0-6c3f352d2257\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://627cd6b39fdedd967fffcfc0755439277b2d73016fba1ddc7f5b9d5deba43b8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p5swq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-t4fq2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:13Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.510681 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rq622" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a742e7b-c420-46e3-9e96-e9c744af6124\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://afeb20035224feeab28a92ac77b43a24e653e49c56a25590a9861019a2b7a8ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad7222c9b91c0065a545bc1904d9864a5923cc13bfb6617daeb4a965a830f191\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:59:45Z\\\",\\\"message\\\":\\\"2026-01-26T16:59:00+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_41ab0694-d9c8-49a7-bf30-57e732ac7550\\\\n2026-01-26T16:59:00+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_41ab0694-d9c8-49a7-bf30-57e732ac7550 to /host/opt/cni/bin/\\\\n2026-01-26T16:59:00Z [verbose] multus-daemon started\\\\n2026-01-26T16:59:00Z [verbose] Readiness Indicator file check\\\\n2026-01-26T16:59:45Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8plh8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rq622\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:13Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.530016 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59ecd87a-c5db-446d-ad3e-cfabbd648c1d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://89975b8f9428f81ab5d3fb48ced5dd9c837bea2feea3b89f5f7ff8d7d5d15b3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://40d432159a07ba20bf95f058bf8a597d67cbd8d852519bed035694f8ba3d8ec4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://59ba0f131eee14df389391e46bc772894e49e976c3663df5fb3bc98b3cb4a3d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03a01b651d9f66da4dd1f6e0d29ad97c0d6ae46b644c3d997d8dc99476706df\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f07438e20bdf71c752bb661084c835341999c561bb5442d75e177223881276f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T16:58:51Z\\\",\\\"message\\\":\\\"falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 16:58:43.431198 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 16:58:43.432227 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-99450334/tls.crt::/tmp/serving-cert-99450334/tls.key\\\\\\\"\\\\nI0126 16:58:50.828320 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 16:58:50.830449 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 16:58:50.830472 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 16:58:50.830509 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 16:58:50.830515 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 16:58:50.834885 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 16:58:50.834958 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 16:58:50.834991 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI0126 16:58:50.834905 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0126 16:58:50.835015 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 16:58:50.835084 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 16:58:50.835106 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 16:58:50.835116 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0126 16:58:50.837641 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:32Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2daa3f13d37b7ae19dfca406b6b5cfdcd4f211287f7d410193f2fee36a24553\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:30Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:13Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.534388 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.534475 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.534520 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.534602 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.534662 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:13Z","lastTransitionTime":"2026-01-26T17:00:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.552420 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab14b0d1-ba6c-4e70-bc80-f4364577742a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a00494ca589263eb0f50c879c0aa1e1c263f74e302325f88eee31b220ebf53b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb3c5348b8b83991cbb42255dc07d74fe50e200793efe1a7b2b2727a5c2be800\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7c3027fabe8a104141386b9767218f38a143318580dd2a33448fed2c05688ba1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a03e2fad94ce4122f1d77ce30dc80bb78298396649c12b885c386e5f8eea50b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:13Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.556460 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 12:56:45.815776686 +0000 UTC Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.579321 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://203d756903498bed1c57a3e87a95f4b24f808514567a99505ba2f3cfec468cb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://203d756903498bed1c57a3e87a95f4b24f808514567a99505ba2f3cfec468cb6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T16:59:58Z\\\",\\\"message\\\":\\\" server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0126 16:59:57.729092 6886 port_cache.go:96] port-cache(openshift-network-diagnostics_network-check-target-xd92c): added port \\\\u0026{name:openshift-network-diagnostics_network-check-target-xd92c uuid:61897e97-c771-4738-8709-09636387cb00 logicalSwitch:crc ips:[0xc008f54210] mac:[10 88 10 217 0 4] expires:{wall:0 ext:0 loc:\\\\u003cnil\\\\u003e}} with IP: [10.217.0.4/23] and MAC: 0a:58:0a:d9:00:04\\\\nI0126 16:59:57.729124 6886 pods.go:252] [openshift-network-diagnostics/network-check-target-xd92c] addLogicalPort took 1.706822ms, libovsdb time 885.597µs\\\\nI0126 16:59:57.729132 6886 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-diagnostics/network-check-target-xd92c after 0 failed attempt(s)\\\\nF0126 16:59:57.729134 6886 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": fa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pxh94_openshift-ovn-kubernetes(ab5b6f50-172b-4535-a0f9-5d103bcab4e7)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9kdbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pxh94\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:13Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.597360 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-295wr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12e50462-28e6-4531-ada4-e652310e6cce\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tf98h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-295wr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:13Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.611070 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:13Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.631257 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad7b59f9-beb7-49d6-a2d1-e29133e46854\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc53a6a6d4222ee5e1ac29fd5957d8cdf3fd42de72c68c85329374ce7afc4004\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e19d1ef39aec00337edf433f38ed92c785d6dfeb404363670b4740889fe00363\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fd0aff003d65fa0a8f2caa68f9024240a3cd07f45721145057419ae07d30b196\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79548fe5ee0a5b0251f7fa0304f57efb25ce63c54e5bb06a0da69e483461d993\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0c0b092f1edef091b4888c2231fd7ffa210e7233355b50d43cba49ffcb897d1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://62b76bf7fd89103159957c0550c2cf0b2dd3dba63dc4d74c6f8acc845ba1edea\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://249b3a1717ed5f1cca8779359055f1d0e36eb8dbb8c5afba0b91d36ebd571da8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:59:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:59:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zm9x6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-v2l7v\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:13Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.637130 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.637213 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.637224 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.637243 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.637257 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:13Z","lastTransitionTime":"2026-01-26T17:00:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.648431 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63c75ede-5170-4db0-811b-5217ef8d72b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da26ddcfc6ccde3c9aabd63bdef3435f9a9eaab8c095bfeef5670c1295576cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54ca3fa13d9e8d442efa93b44a870369f7df3fe7562d77b98528f5c19a751f18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96lw2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xm9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:13Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.669437 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc864b0d-83bc-4954-9c61-ad650157caff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cbde934a6c8acad10ca3ab8206d0ddbd4f7b17e9d304b898a68f4d3b0303bb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b48f763d4aff37169399be766d5ab4f7ebbf91f304d139c9022a8556946eb107\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0094e662f53c4832a984e05a880021af05ffc4c27f25394c28a070d9ef5490d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cb95f623fa4f5f42217649ccce4225f9fe588bb3558da9334eef2b40ddb62486\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb95f623fa4f5f42217649ccce4225f9fe588bb3558da9334eef2b40ddb62486\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:13Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.691738 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2c3b0574-b4cc-483d-ae88-6517d1f30772\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec9063a7c03990fc26fc47427f164a769fd649c2bdbd9d23ea7f646e569734be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f67797295a8c3952902b2696c6fdb26b72ce1826b5ccd522a24aac90a0411b5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f67797295a8c3952902b2696c6fdb26b72ce1826b5ccd522a24aac90a0411b5f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:13Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.710387 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c09748c0963bdb50c9552f249d1135ea53e68c52babd788dad050951cb849cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:13Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.727691 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:13Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.739920 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.739952 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.739966 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.739986 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.740000 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:13Z","lastTransitionTime":"2026-01-26T17:00:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.743385 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v7579" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a77e85f9-b566-4807-bb92-55963c97b93c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:59:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba87c9fc35c230bbee201a5176cb467309f0b9aee82dfc81f3b677a15486d02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c03dc794e9c2035f2e1983eacad3e51d76223cb1b82e2f402c73f9453e4bd2f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:59:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4n9h7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:59:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-v7579\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:13Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.770666 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"280cd8ed-5dcf-487a-8b00-22204e94d54f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c387234ad8d7123da333d3de4a80f3a79c25dddf0c3a0fb004b521161ff105b4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d65b21fc101230cb18ee921fc481e83c944dde8fe01074931b90551e082ee249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff371891a210c6f3498b0d8377c477749a9ea438aa74f1e33f8ac9047df447ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c687b137e2bdfb70b19588ae8f5c65a23c2df57716cfd6918856236f2d6610a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93c814433ba35046d47c29524f19b728793436e9f6967a6ea7249e35f673f48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bf39fbfd0b23f9b34e42610ae3603d849bcf4211f53ba47cbbebdaf47a9687d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bf39fbfd0b23f9b34e42610ae3603d849bcf4211f53ba47cbbebdaf47a9687d8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2ca7ee60b82663fdc02dc2dd3f7af379df8407800d04c57f4f4d09d49ed9aa0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a2ca7ee60b82663fdc02dc2dd3f7af379df8407800d04c57f4f4d09d49ed9aa0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:28Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://ee2a878cbd2cdef8fe8d9bb62a4554ffb8aeadfb90ab92b4ff6ec965824ec37a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ee2a878cbd2cdef8fe8d9bb62a4554ffb8aeadfb90ab92b4ff6ec965824ec37a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T16:58:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T16:58:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:25Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:13Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.787653 4856 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-tp5hk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f28414c-12c1-4adb-be7b-6182310828eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T16:58:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0af4930ff9cebd62106545fb3da2dfc93f7b591426fe4c85aa2e637b60c935f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T16:58:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzc59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T16:58:54Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-tp5hk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T17:00:13Z is after 2025-08-24T17:21:41Z" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.843374 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.843441 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.843465 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.843492 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.843508 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:13Z","lastTransitionTime":"2026-01-26T17:00:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.946393 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.946474 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.946495 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.946577 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:13 crc kubenswrapper[4856]: I0126 17:00:13.946607 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:13Z","lastTransitionTime":"2026-01-26T17:00:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.050309 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.050370 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.050389 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.050416 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.050435 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:14Z","lastTransitionTime":"2026-01-26T17:00:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.154041 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.154122 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.154140 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.154166 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.154184 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:14Z","lastTransitionTime":"2026-01-26T17:00:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.258026 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.258078 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.258096 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.258120 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.258139 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:14Z","lastTransitionTime":"2026-01-26T17:00:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.360139 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.360188 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.360206 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.360230 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.360248 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:14Z","lastTransitionTime":"2026-01-26T17:00:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.378940 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12e50462-28e6-4531-ada4-e652310e6cce-metrics-certs\") pod \"network-metrics-daemon-295wr\" (UID: \"12e50462-28e6-4531-ada4-e652310e6cce\") " pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 17:00:14 crc kubenswrapper[4856]: E0126 17:00:14.379207 4856 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 17:00:14 crc kubenswrapper[4856]: E0126 17:00:14.379382 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12e50462-28e6-4531-ada4-e652310e6cce-metrics-certs podName:12e50462-28e6-4531-ada4-e652310e6cce nodeName:}" failed. No retries permitted until 2026-01-26 17:01:18.37933313 +0000 UTC m=+174.332587121 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/12e50462-28e6-4531-ada4-e652310e6cce-metrics-certs") pod "network-metrics-daemon-295wr" (UID: "12e50462-28e6-4531-ada4-e652310e6cce") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.394935 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.394988 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.395058 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.395089 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 17:00:14 crc kubenswrapper[4856]: E0126 17:00:14.395379 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 17:00:14 crc kubenswrapper[4856]: E0126 17:00:14.395593 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 17:00:14 crc kubenswrapper[4856]: E0126 17:00:14.395797 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 17:00:14 crc kubenswrapper[4856]: E0126 17:00:14.395924 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.462199 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.462236 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.462247 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.462262 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.462273 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:14Z","lastTransitionTime":"2026-01-26T17:00:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.556576 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 08:38:11.859416206 +0000 UTC Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.564496 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.564558 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.564571 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.564585 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.564596 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:14Z","lastTransitionTime":"2026-01-26T17:00:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.791270 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.791304 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.791312 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.791328 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.791340 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:14Z","lastTransitionTime":"2026-01-26T17:00:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.893462 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.893510 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.893522 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.893563 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.893575 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:14Z","lastTransitionTime":"2026-01-26T17:00:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.997413 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.997496 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.997554 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.997592 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:14 crc kubenswrapper[4856]: I0126 17:00:14.997619 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:14Z","lastTransitionTime":"2026-01-26T17:00:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.100460 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.100821 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.100996 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.101149 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.101275 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:15Z","lastTransitionTime":"2026-01-26T17:00:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.203905 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.204189 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.204257 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.204323 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.204379 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:15Z","lastTransitionTime":"2026-01-26T17:00:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.307481 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.307555 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.307564 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.307582 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.307591 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:15Z","lastTransitionTime":"2026-01-26T17:00:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.411958 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.412010 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.412022 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.412039 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.412054 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:15Z","lastTransitionTime":"2026-01-26T17:00:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.568400 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 14:43:38.756110219 +0000 UTC Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.571088 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.571154 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.571165 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.571196 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.571209 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:15Z","lastTransitionTime":"2026-01-26T17:00:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.597386 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-t4fq2" podStartSLOduration=82.597346293 podStartE2EDuration="1m22.597346293s" podCreationTimestamp="2026-01-26 16:58:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:00:15.582847484 +0000 UTC m=+111.536101465" watchObservedRunningTime="2026-01-26 17:00:15.597346293 +0000 UTC m=+111.550600284" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.597899 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-rq622" podStartSLOduration=81.59789124 podStartE2EDuration="1m21.59789124s" podCreationTimestamp="2026-01-26 16:58:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:00:15.597551549 +0000 UTC m=+111.550805540" watchObservedRunningTime="2026-01-26 17:00:15.59789124 +0000 UTC m=+111.551145221" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.622576 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=67.622560916 podStartE2EDuration="1m7.622560916s" podCreationTimestamp="2026-01-26 16:59:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:00:15.622380181 +0000 UTC m=+111.575634182" watchObservedRunningTime="2026-01-26 17:00:15.622560916 +0000 UTC m=+111.575814897" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.642950 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=7.642924933 podStartE2EDuration="7.642924933s" podCreationTimestamp="2026-01-26 17:00:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:00:15.642045726 +0000 UTC m=+111.595299707" watchObservedRunningTime="2026-01-26 17:00:15.642924933 +0000 UTC m=+111.596178924" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.674183 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.674224 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.674237 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.674252 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.674264 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:15Z","lastTransitionTime":"2026-01-26T17:00:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.720910 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-v2l7v" podStartSLOduration=81.720889012 podStartE2EDuration="1m21.720889012s" podCreationTimestamp="2026-01-26 16:58:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:00:15.720592533 +0000 UTC m=+111.673846524" watchObservedRunningTime="2026-01-26 17:00:15.720889012 +0000 UTC m=+111.674143003" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.752067 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podStartSLOduration=81.752051285 podStartE2EDuration="1m21.752051285s" podCreationTimestamp="2026-01-26 16:58:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:00:15.735637638 +0000 UTC m=+111.688891629" watchObservedRunningTime="2026-01-26 17:00:15.752051285 +0000 UTC m=+111.705305266" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.752243 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=62.752239391 podStartE2EDuration="1m2.752239391s" podCreationTimestamp="2026-01-26 16:59:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:00:15.751911751 +0000 UTC m=+111.705165742" watchObservedRunningTime="2026-01-26 17:00:15.752239391 +0000 UTC m=+111.705493372" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.775643 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=30.775623948 podStartE2EDuration="30.775623948s" podCreationTimestamp="2026-01-26 16:59:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:00:15.762891513 +0000 UTC m=+111.716145494" watchObservedRunningTime="2026-01-26 17:00:15.775623948 +0000 UTC m=+111.728877929" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.776310 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.776370 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.776380 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.776395 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.776405 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:15Z","lastTransitionTime":"2026-01-26T17:00:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.811769 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-v7579" podStartSLOduration=80.811752622 podStartE2EDuration="1m20.811752622s" podCreationTimestamp="2026-01-26 16:58:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:00:15.811435272 +0000 UTC m=+111.764689273" watchObservedRunningTime="2026-01-26 17:00:15.811752622 +0000 UTC m=+111.765006603" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.839850 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=8.839828292 podStartE2EDuration="8.839828292s" podCreationTimestamp="2026-01-26 17:00:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:00:15.837423199 +0000 UTC m=+111.790677190" watchObservedRunningTime="2026-01-26 17:00:15.839828292 +0000 UTC m=+111.793082283" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.853243 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-tp5hk" podStartSLOduration=82.853216757 podStartE2EDuration="1m22.853216757s" podCreationTimestamp="2026-01-26 16:58:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:00:15.852262528 +0000 UTC m=+111.805516529" watchObservedRunningTime="2026-01-26 17:00:15.853216757 +0000 UTC m=+111.806470758" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.879465 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.879507 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.879519 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.879545 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.879554 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:15Z","lastTransitionTime":"2026-01-26T17:00:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.981682 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.981717 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.981726 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.981741 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:15 crc kubenswrapper[4856]: I0126 17:00:15.981769 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:15Z","lastTransitionTime":"2026-01-26T17:00:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.084473 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.084543 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.084555 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.084575 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.084588 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:16Z","lastTransitionTime":"2026-01-26T17:00:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.187061 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.187087 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.187095 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.187107 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.187115 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:16Z","lastTransitionTime":"2026-01-26T17:00:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.289744 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.289802 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.289824 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.289849 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.289866 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:16Z","lastTransitionTime":"2026-01-26T17:00:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.393722 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.393777 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.393792 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.393819 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.393832 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:16Z","lastTransitionTime":"2026-01-26T17:00:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.394179 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.394230 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.394269 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.394344 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 17:00:16 crc kubenswrapper[4856]: E0126 17:00:16.394386 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 17:00:16 crc kubenswrapper[4856]: E0126 17:00:16.394450 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 17:00:16 crc kubenswrapper[4856]: E0126 17:00:16.394591 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 17:00:16 crc kubenswrapper[4856]: E0126 17:00:16.394793 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.496373 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.496407 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.496417 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.496432 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.496442 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:16Z","lastTransitionTime":"2026-01-26T17:00:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.587314 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 15:27:07.353976077 +0000 UTC Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.599773 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.599818 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.599830 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.599852 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.599864 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:16Z","lastTransitionTime":"2026-01-26T17:00:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.704133 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.704207 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.704234 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.704287 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.704311 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:16Z","lastTransitionTime":"2026-01-26T17:00:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.807487 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.807828 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.807840 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.807857 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.807869 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:16Z","lastTransitionTime":"2026-01-26T17:00:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.910367 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.910423 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.910443 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.910468 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:16 crc kubenswrapper[4856]: I0126 17:00:16.910486 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:16Z","lastTransitionTime":"2026-01-26T17:00:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.012496 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.012582 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.012597 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.012622 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.012634 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:17Z","lastTransitionTime":"2026-01-26T17:00:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.116219 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.116259 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.116272 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.116290 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.116302 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:17Z","lastTransitionTime":"2026-01-26T17:00:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.218659 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.218701 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.218714 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.218754 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.218770 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:17Z","lastTransitionTime":"2026-01-26T17:00:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.428243 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.428270 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.428278 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.428289 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.428299 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:17Z","lastTransitionTime":"2026-01-26T17:00:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.530737 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.530779 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.530791 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.530808 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.530820 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:17Z","lastTransitionTime":"2026-01-26T17:00:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.588230 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 18:28:46.994851214 +0000 UTC Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.633223 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.633281 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.633295 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.633316 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.633348 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:17Z","lastTransitionTime":"2026-01-26T17:00:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.736147 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.736213 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.736228 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.736272 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.736287 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:17Z","lastTransitionTime":"2026-01-26T17:00:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.838881 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.838931 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.838943 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.838962 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.838973 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:17Z","lastTransitionTime":"2026-01-26T17:00:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.941431 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.941500 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.941518 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.941598 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:17 crc kubenswrapper[4856]: I0126 17:00:17.941621 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:17Z","lastTransitionTime":"2026-01-26T17:00:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.044545 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.044592 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.044606 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.044623 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.044638 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:18Z","lastTransitionTime":"2026-01-26T17:00:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.147935 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.147998 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.148017 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.148041 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.148059 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:18Z","lastTransitionTime":"2026-01-26T17:00:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.253889 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.253943 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.253960 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.253984 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.253998 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:18Z","lastTransitionTime":"2026-01-26T17:00:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.356691 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.356732 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.356742 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.356757 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.356768 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:18Z","lastTransitionTime":"2026-01-26T17:00:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.395107 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.395170 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.395270 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.395346 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 17:00:18 crc kubenswrapper[4856]: E0126 17:00:18.395673 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 17:00:18 crc kubenswrapper[4856]: E0126 17:00:18.395857 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 17:00:18 crc kubenswrapper[4856]: E0126 17:00:18.396009 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 17:00:18 crc kubenswrapper[4856]: E0126 17:00:18.396135 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.459839 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.459895 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.459913 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.459934 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.459947 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:18Z","lastTransitionTime":"2026-01-26T17:00:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.563076 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.563112 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.563125 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.563176 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.563188 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:18Z","lastTransitionTime":"2026-01-26T17:00:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.588589 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 22:36:58.335702397 +0000 UTC Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.666269 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.666309 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.666321 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.666337 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.666346 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:18Z","lastTransitionTime":"2026-01-26T17:00:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.776869 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.776940 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.776955 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.776974 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.776989 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:18Z","lastTransitionTime":"2026-01-26T17:00:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.880464 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.880502 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.880511 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.880546 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.880558 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:18Z","lastTransitionTime":"2026-01-26T17:00:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.984139 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.984204 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.984221 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.984246 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:18 crc kubenswrapper[4856]: I0126 17:00:18.984266 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:18Z","lastTransitionTime":"2026-01-26T17:00:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:19 crc kubenswrapper[4856]: I0126 17:00:19.087342 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:19 crc kubenswrapper[4856]: I0126 17:00:19.087386 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:19 crc kubenswrapper[4856]: I0126 17:00:19.087395 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:19 crc kubenswrapper[4856]: I0126 17:00:19.087411 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:19 crc kubenswrapper[4856]: I0126 17:00:19.087422 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:19Z","lastTransitionTime":"2026-01-26T17:00:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:19 crc kubenswrapper[4856]: I0126 17:00:19.189991 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:19 crc kubenswrapper[4856]: I0126 17:00:19.190035 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:19 crc kubenswrapper[4856]: I0126 17:00:19.190044 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:19 crc kubenswrapper[4856]: I0126 17:00:19.190061 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:19 crc kubenswrapper[4856]: I0126 17:00:19.190070 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:19Z","lastTransitionTime":"2026-01-26T17:00:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:19 crc kubenswrapper[4856]: I0126 17:00:19.293634 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:19 crc kubenswrapper[4856]: I0126 17:00:19.293672 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:19 crc kubenswrapper[4856]: I0126 17:00:19.293683 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:19 crc kubenswrapper[4856]: I0126 17:00:19.293713 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:19 crc kubenswrapper[4856]: I0126 17:00:19.293724 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:19Z","lastTransitionTime":"2026-01-26T17:00:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:19 crc kubenswrapper[4856]: I0126 17:00:19.485208 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:19 crc kubenswrapper[4856]: I0126 17:00:19.485244 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:19 crc kubenswrapper[4856]: I0126 17:00:19.485255 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:19 crc kubenswrapper[4856]: I0126 17:00:19.485273 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:19 crc kubenswrapper[4856]: I0126 17:00:19.485285 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:19Z","lastTransitionTime":"2026-01-26T17:00:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:19 crc kubenswrapper[4856]: I0126 17:00:19.588384 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:19 crc kubenswrapper[4856]: I0126 17:00:19.588450 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:19 crc kubenswrapper[4856]: I0126 17:00:19.588461 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:19 crc kubenswrapper[4856]: I0126 17:00:19.588479 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:19 crc kubenswrapper[4856]: I0126 17:00:19.588490 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:19Z","lastTransitionTime":"2026-01-26T17:00:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:19 crc kubenswrapper[4856]: I0126 17:00:19.588701 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 08:23:55.898464011 +0000 UTC Jan 26 17:00:19 crc kubenswrapper[4856]: I0126 17:00:19.697128 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:19 crc kubenswrapper[4856]: I0126 17:00:19.697179 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:19 crc kubenswrapper[4856]: I0126 17:00:19.697193 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:19 crc kubenswrapper[4856]: I0126 17:00:19.697215 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:19 crc kubenswrapper[4856]: I0126 17:00:19.697231 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:19Z","lastTransitionTime":"2026-01-26T17:00:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:19 crc kubenswrapper[4856]: I0126 17:00:19.799263 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:19 crc kubenswrapper[4856]: I0126 17:00:19.799319 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:19 crc kubenswrapper[4856]: I0126 17:00:19.799328 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:19 crc kubenswrapper[4856]: I0126 17:00:19.799340 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:19 crc kubenswrapper[4856]: I0126 17:00:19.799349 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:19Z","lastTransitionTime":"2026-01-26T17:00:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:19 crc kubenswrapper[4856]: I0126 17:00:19.901637 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:19 crc kubenswrapper[4856]: I0126 17:00:19.901666 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:19 crc kubenswrapper[4856]: I0126 17:00:19.901676 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:19 crc kubenswrapper[4856]: I0126 17:00:19.901692 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:19 crc kubenswrapper[4856]: I0126 17:00:19.901702 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:19Z","lastTransitionTime":"2026-01-26T17:00:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.003840 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.003880 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.003892 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.003909 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.003921 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:20Z","lastTransitionTime":"2026-01-26T17:00:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.106968 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.107016 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.107027 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.107043 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.107055 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:20Z","lastTransitionTime":"2026-01-26T17:00:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.209869 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.209933 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.209947 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.209971 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.209988 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:20Z","lastTransitionTime":"2026-01-26T17:00:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.312466 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.312508 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.312517 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.312556 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.312565 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:20Z","lastTransitionTime":"2026-01-26T17:00:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.394742 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.394848 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 17:00:20 crc kubenswrapper[4856]: E0126 17:00:20.394910 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.394978 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.395002 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 17:00:20 crc kubenswrapper[4856]: E0126 17:00:20.395203 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 17:00:20 crc kubenswrapper[4856]: E0126 17:00:20.395315 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 17:00:20 crc kubenswrapper[4856]: E0126 17:00:20.395411 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.416274 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.416339 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.416355 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.416378 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.416393 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:20Z","lastTransitionTime":"2026-01-26T17:00:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.519913 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.519982 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.519997 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.520018 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.520030 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:20Z","lastTransitionTime":"2026-01-26T17:00:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.579580 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.579815 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.579898 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.579947 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.579976 4856 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T17:00:20Z","lastTransitionTime":"2026-01-26T17:00:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.589762 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 10:45:44.864025374 +0000 UTC Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.658955 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-fdsb9"] Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.659716 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fdsb9" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.662516 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.662641 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.663557 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.665290 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.695394 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e7ec14e-9348-48ea-ae7c-5cf3974b7a55-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-fdsb9\" (UID: \"7e7ec14e-9348-48ea-ae7c-5cf3974b7a55\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fdsb9" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.695431 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7e7ec14e-9348-48ea-ae7c-5cf3974b7a55-service-ca\") pod \"cluster-version-operator-5c965bbfc6-fdsb9\" (UID: \"7e7ec14e-9348-48ea-ae7c-5cf3974b7a55\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fdsb9" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.695478 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7e7ec14e-9348-48ea-ae7c-5cf3974b7a55-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-fdsb9\" (UID: \"7e7ec14e-9348-48ea-ae7c-5cf3974b7a55\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fdsb9" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.695514 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/7e7ec14e-9348-48ea-ae7c-5cf3974b7a55-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-fdsb9\" (UID: \"7e7ec14e-9348-48ea-ae7c-5cf3974b7a55\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fdsb9" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.695574 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/7e7ec14e-9348-48ea-ae7c-5cf3974b7a55-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-fdsb9\" (UID: \"7e7ec14e-9348-48ea-ae7c-5cf3974b7a55\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fdsb9" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.796568 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7e7ec14e-9348-48ea-ae7c-5cf3974b7a55-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-fdsb9\" (UID: \"7e7ec14e-9348-48ea-ae7c-5cf3974b7a55\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fdsb9" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.796668 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/7e7ec14e-9348-48ea-ae7c-5cf3974b7a55-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-fdsb9\" (UID: \"7e7ec14e-9348-48ea-ae7c-5cf3974b7a55\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fdsb9" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.796729 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/7e7ec14e-9348-48ea-ae7c-5cf3974b7a55-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-fdsb9\" (UID: \"7e7ec14e-9348-48ea-ae7c-5cf3974b7a55\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fdsb9" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.796820 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e7ec14e-9348-48ea-ae7c-5cf3974b7a55-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-fdsb9\" (UID: \"7e7ec14e-9348-48ea-ae7c-5cf3974b7a55\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fdsb9" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.796859 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7e7ec14e-9348-48ea-ae7c-5cf3974b7a55-service-ca\") pod \"cluster-version-operator-5c965bbfc6-fdsb9\" (UID: \"7e7ec14e-9348-48ea-ae7c-5cf3974b7a55\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fdsb9" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.796904 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/7e7ec14e-9348-48ea-ae7c-5cf3974b7a55-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-fdsb9\" (UID: \"7e7ec14e-9348-48ea-ae7c-5cf3974b7a55\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fdsb9" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.796971 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/7e7ec14e-9348-48ea-ae7c-5cf3974b7a55-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-fdsb9\" (UID: \"7e7ec14e-9348-48ea-ae7c-5cf3974b7a55\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fdsb9" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.798136 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7e7ec14e-9348-48ea-ae7c-5cf3974b7a55-service-ca\") pod \"cluster-version-operator-5c965bbfc6-fdsb9\" (UID: \"7e7ec14e-9348-48ea-ae7c-5cf3974b7a55\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fdsb9" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.804212 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e7ec14e-9348-48ea-ae7c-5cf3974b7a55-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-fdsb9\" (UID: \"7e7ec14e-9348-48ea-ae7c-5cf3974b7a55\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fdsb9" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.828738 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7e7ec14e-9348-48ea-ae7c-5cf3974b7a55-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-fdsb9\" (UID: \"7e7ec14e-9348-48ea-ae7c-5cf3974b7a55\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fdsb9" Jan 26 17:00:20 crc kubenswrapper[4856]: I0126 17:00:20.976275 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fdsb9" Jan 26 17:00:21 crc kubenswrapper[4856]: W0126 17:00:21.005387 4856 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e7ec14e_9348_48ea_ae7c_5cf3974b7a55.slice/crio-79b8671c2619e7cea07cffb5d2d7bd4c1ea0c3444ee981556807083a2163c60a WatchSource:0}: Error finding container 79b8671c2619e7cea07cffb5d2d7bd4c1ea0c3444ee981556807083a2163c60a: Status 404 returned error can't find the container with id 79b8671c2619e7cea07cffb5d2d7bd4c1ea0c3444ee981556807083a2163c60a Jan 26 17:00:21 crc kubenswrapper[4856]: I0126 17:00:21.590618 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 18:01:47.884230807 +0000 UTC Jan 26 17:00:21 crc kubenswrapper[4856]: I0126 17:00:21.590916 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Jan 26 17:00:21 crc kubenswrapper[4856]: I0126 17:00:21.605120 4856 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Jan 26 17:00:21 crc kubenswrapper[4856]: I0126 17:00:21.963192 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fdsb9" event={"ID":"7e7ec14e-9348-48ea-ae7c-5cf3974b7a55","Type":"ContainerStarted","Data":"64170a572cc954a40ed5bf4f9d7a0fff4dd4a74e7c0e71b32b75b571217d368a"} Jan 26 17:00:21 crc kubenswrapper[4856]: I0126 17:00:21.963279 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fdsb9" event={"ID":"7e7ec14e-9348-48ea-ae7c-5cf3974b7a55","Type":"ContainerStarted","Data":"79b8671c2619e7cea07cffb5d2d7bd4c1ea0c3444ee981556807083a2163c60a"} Jan 26 17:00:22 crc kubenswrapper[4856]: I0126 17:00:22.394392 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 17:00:22 crc kubenswrapper[4856]: I0126 17:00:22.394392 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 17:00:22 crc kubenswrapper[4856]: I0126 17:00:22.394576 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 17:00:22 crc kubenswrapper[4856]: I0126 17:00:22.394892 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 17:00:22 crc kubenswrapper[4856]: E0126 17:00:22.395089 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 17:00:22 crc kubenswrapper[4856]: E0126 17:00:22.395203 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 17:00:22 crc kubenswrapper[4856]: E0126 17:00:22.395289 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 17:00:22 crc kubenswrapper[4856]: E0126 17:00:22.395452 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 17:00:24 crc kubenswrapper[4856]: I0126 17:00:24.394224 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 17:00:24 crc kubenswrapper[4856]: E0126 17:00:24.394349 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 17:00:24 crc kubenswrapper[4856]: I0126 17:00:24.394687 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 17:00:24 crc kubenswrapper[4856]: I0126 17:00:24.394718 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 17:00:24 crc kubenswrapper[4856]: E0126 17:00:24.394766 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 17:00:24 crc kubenswrapper[4856]: I0126 17:00:24.394791 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 17:00:24 crc kubenswrapper[4856]: E0126 17:00:24.394961 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 17:00:24 crc kubenswrapper[4856]: I0126 17:00:24.395107 4856 scope.go:117] "RemoveContainer" containerID="203d756903498bed1c57a3e87a95f4b24f808514567a99505ba2f3cfec468cb6" Jan 26 17:00:24 crc kubenswrapper[4856]: E0126 17:00:24.395249 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pxh94_openshift-ovn-kubernetes(ab5b6f50-172b-4535-a0f9-5d103bcab4e7)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" Jan 26 17:00:24 crc kubenswrapper[4856]: E0126 17:00:24.395300 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 17:00:25 crc kubenswrapper[4856]: E0126 17:00:25.322644 4856 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Jan 26 17:00:25 crc kubenswrapper[4856]: E0126 17:00:25.732538 4856 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 26 17:00:26 crc kubenswrapper[4856]: I0126 17:00:26.394913 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 17:00:26 crc kubenswrapper[4856]: I0126 17:00:26.394952 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 17:00:26 crc kubenswrapper[4856]: I0126 17:00:26.394999 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 17:00:26 crc kubenswrapper[4856]: E0126 17:00:26.395262 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 17:00:26 crc kubenswrapper[4856]: E0126 17:00:26.395370 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 17:00:26 crc kubenswrapper[4856]: E0126 17:00:26.395412 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 17:00:26 crc kubenswrapper[4856]: I0126 17:00:26.395576 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 17:00:26 crc kubenswrapper[4856]: E0126 17:00:26.395642 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 17:00:28 crc kubenswrapper[4856]: I0126 17:00:28.395174 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 17:00:28 crc kubenswrapper[4856]: I0126 17:00:28.395248 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 17:00:28 crc kubenswrapper[4856]: I0126 17:00:28.395337 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 17:00:28 crc kubenswrapper[4856]: I0126 17:00:28.395393 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 17:00:28 crc kubenswrapper[4856]: E0126 17:00:28.395962 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 17:00:28 crc kubenswrapper[4856]: E0126 17:00:28.396405 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 17:00:28 crc kubenswrapper[4856]: E0126 17:00:28.396272 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 17:00:28 crc kubenswrapper[4856]: E0126 17:00:28.396617 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 17:00:30 crc kubenswrapper[4856]: I0126 17:00:30.394791 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 17:00:30 crc kubenswrapper[4856]: I0126 17:00:30.394824 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 17:00:30 crc kubenswrapper[4856]: I0126 17:00:30.394824 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 17:00:30 crc kubenswrapper[4856]: I0126 17:00:30.394865 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 17:00:30 crc kubenswrapper[4856]: E0126 17:00:30.395807 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 17:00:30 crc kubenswrapper[4856]: E0126 17:00:30.395955 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 17:00:30 crc kubenswrapper[4856]: E0126 17:00:30.396205 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 17:00:30 crc kubenswrapper[4856]: E0126 17:00:30.396244 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 17:00:30 crc kubenswrapper[4856]: E0126 17:00:30.733490 4856 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 26 17:00:32 crc kubenswrapper[4856]: I0126 17:00:32.395068 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 17:00:32 crc kubenswrapper[4856]: I0126 17:00:32.395116 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 17:00:32 crc kubenswrapper[4856]: I0126 17:00:32.395155 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 17:00:32 crc kubenswrapper[4856]: E0126 17:00:32.395241 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 17:00:32 crc kubenswrapper[4856]: I0126 17:00:32.395276 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 17:00:32 crc kubenswrapper[4856]: E0126 17:00:32.395503 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 17:00:32 crc kubenswrapper[4856]: E0126 17:00:32.395538 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 17:00:32 crc kubenswrapper[4856]: E0126 17:00:32.395623 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 17:00:33 crc kubenswrapper[4856]: I0126 17:00:33.003342 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-rq622_7a742e7b-c420-46e3-9e96-e9c744af6124/kube-multus/1.log" Jan 26 17:00:33 crc kubenswrapper[4856]: I0126 17:00:33.004282 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-rq622_7a742e7b-c420-46e3-9e96-e9c744af6124/kube-multus/0.log" Jan 26 17:00:33 crc kubenswrapper[4856]: I0126 17:00:33.004356 4856 generic.go:334] "Generic (PLEG): container finished" podID="7a742e7b-c420-46e3-9e96-e9c744af6124" containerID="afeb20035224feeab28a92ac77b43a24e653e49c56a25590a9861019a2b7a8ff" exitCode=1 Jan 26 17:00:33 crc kubenswrapper[4856]: I0126 17:00:33.004398 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-rq622" event={"ID":"7a742e7b-c420-46e3-9e96-e9c744af6124","Type":"ContainerDied","Data":"afeb20035224feeab28a92ac77b43a24e653e49c56a25590a9861019a2b7a8ff"} Jan 26 17:00:33 crc kubenswrapper[4856]: I0126 17:00:33.004452 4856 scope.go:117] "RemoveContainer" containerID="ad7222c9b91c0065a545bc1904d9864a5923cc13bfb6617daeb4a965a830f191" Jan 26 17:00:33 crc kubenswrapper[4856]: I0126 17:00:33.004882 4856 scope.go:117] "RemoveContainer" containerID="afeb20035224feeab28a92ac77b43a24e653e49c56a25590a9861019a2b7a8ff" Jan 26 17:00:33 crc kubenswrapper[4856]: E0126 17:00:33.005055 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-rq622_openshift-multus(7a742e7b-c420-46e3-9e96-e9c744af6124)\"" pod="openshift-multus/multus-rq622" podUID="7a742e7b-c420-46e3-9e96-e9c744af6124" Jan 26 17:00:33 crc kubenswrapper[4856]: I0126 17:00:33.026639 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fdsb9" podStartSLOduration=99.026617703 podStartE2EDuration="1m39.026617703s" podCreationTimestamp="2026-01-26 16:58:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:00:21.989087162 +0000 UTC m=+117.942341163" watchObservedRunningTime="2026-01-26 17:00:33.026617703 +0000 UTC m=+128.979871684" Jan 26 17:00:34 crc kubenswrapper[4856]: I0126 17:00:34.009026 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-rq622_7a742e7b-c420-46e3-9e96-e9c744af6124/kube-multus/1.log" Jan 26 17:00:34 crc kubenswrapper[4856]: I0126 17:00:34.395274 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 17:00:34 crc kubenswrapper[4856]: I0126 17:00:34.395376 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 17:00:34 crc kubenswrapper[4856]: I0126 17:00:34.395386 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 17:00:34 crc kubenswrapper[4856]: I0126 17:00:34.395298 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 17:00:34 crc kubenswrapper[4856]: E0126 17:00:34.395472 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 17:00:34 crc kubenswrapper[4856]: E0126 17:00:34.395620 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 17:00:34 crc kubenswrapper[4856]: E0126 17:00:34.395729 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 17:00:34 crc kubenswrapper[4856]: E0126 17:00:34.395861 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 17:00:35 crc kubenswrapper[4856]: E0126 17:00:35.734853 4856 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 26 17:00:36 crc kubenswrapper[4856]: I0126 17:00:36.395177 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 17:00:36 crc kubenswrapper[4856]: I0126 17:00:36.395225 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 17:00:36 crc kubenswrapper[4856]: I0126 17:00:36.395235 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 17:00:36 crc kubenswrapper[4856]: E0126 17:00:36.395328 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 17:00:36 crc kubenswrapper[4856]: I0126 17:00:36.395351 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 17:00:36 crc kubenswrapper[4856]: E0126 17:00:36.395446 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 17:00:36 crc kubenswrapper[4856]: E0126 17:00:36.395630 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 17:00:36 crc kubenswrapper[4856]: E0126 17:00:36.395702 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 17:00:38 crc kubenswrapper[4856]: I0126 17:00:38.394708 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 17:00:38 crc kubenswrapper[4856]: I0126 17:00:38.394755 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 17:00:38 crc kubenswrapper[4856]: I0126 17:00:38.394997 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 17:00:38 crc kubenswrapper[4856]: I0126 17:00:38.395002 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 17:00:38 crc kubenswrapper[4856]: E0126 17:00:38.395138 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 17:00:38 crc kubenswrapper[4856]: E0126 17:00:38.395280 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 17:00:38 crc kubenswrapper[4856]: E0126 17:00:38.395434 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 17:00:38 crc kubenswrapper[4856]: E0126 17:00:38.395547 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 17:00:39 crc kubenswrapper[4856]: I0126 17:00:39.395942 4856 scope.go:117] "RemoveContainer" containerID="203d756903498bed1c57a3e87a95f4b24f808514567a99505ba2f3cfec468cb6" Jan 26 17:00:40 crc kubenswrapper[4856]: I0126 17:00:40.028761 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pxh94_ab5b6f50-172b-4535-a0f9-5d103bcab4e7/ovnkube-controller/3.log" Jan 26 17:00:40 crc kubenswrapper[4856]: I0126 17:00:40.031560 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" event={"ID":"ab5b6f50-172b-4535-a0f9-5d103bcab4e7","Type":"ContainerStarted","Data":"984517fa62d3e73cf4b33b7c4a101f8221c940f66938120373b7d41cac9c5e5c"} Jan 26 17:00:40 crc kubenswrapper[4856]: I0126 17:00:40.032077 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 17:00:40 crc kubenswrapper[4856]: I0126 17:00:40.164107 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" podStartSLOduration=106.164081989 podStartE2EDuration="1m46.164081989s" podCreationTimestamp="2026-01-26 16:58:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:00:40.065119811 +0000 UTC m=+136.018373802" watchObservedRunningTime="2026-01-26 17:00:40.164081989 +0000 UTC m=+136.117335970" Jan 26 17:00:40 crc kubenswrapper[4856]: I0126 17:00:40.165496 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-295wr"] Jan 26 17:00:40 crc kubenswrapper[4856]: I0126 17:00:40.165615 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 17:00:40 crc kubenswrapper[4856]: E0126 17:00:40.165714 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 17:00:40 crc kubenswrapper[4856]: I0126 17:00:40.394491 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 17:00:40 crc kubenswrapper[4856]: I0126 17:00:40.394556 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 17:00:40 crc kubenswrapper[4856]: E0126 17:00:40.394664 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 17:00:40 crc kubenswrapper[4856]: I0126 17:00:40.394467 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 17:00:40 crc kubenswrapper[4856]: E0126 17:00:40.394798 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 17:00:40 crc kubenswrapper[4856]: E0126 17:00:40.395261 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 17:00:40 crc kubenswrapper[4856]: E0126 17:00:40.736734 4856 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 26 17:00:42 crc kubenswrapper[4856]: I0126 17:00:42.394731 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 17:00:42 crc kubenswrapper[4856]: I0126 17:00:42.394739 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 17:00:42 crc kubenswrapper[4856]: E0126 17:00:42.394895 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 17:00:42 crc kubenswrapper[4856]: I0126 17:00:42.394758 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 17:00:42 crc kubenswrapper[4856]: I0126 17:00:42.394746 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 17:00:42 crc kubenswrapper[4856]: E0126 17:00:42.394985 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 17:00:42 crc kubenswrapper[4856]: E0126 17:00:42.395099 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 17:00:42 crc kubenswrapper[4856]: E0126 17:00:42.395222 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 17:00:44 crc kubenswrapper[4856]: I0126 17:00:44.394196 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 17:00:44 crc kubenswrapper[4856]: I0126 17:00:44.394265 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 17:00:44 crc kubenswrapper[4856]: I0126 17:00:44.394204 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 17:00:44 crc kubenswrapper[4856]: E0126 17:00:44.394363 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 17:00:44 crc kubenswrapper[4856]: I0126 17:00:44.394283 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 17:00:44 crc kubenswrapper[4856]: E0126 17:00:44.394462 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 17:00:44 crc kubenswrapper[4856]: E0126 17:00:44.394602 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 17:00:44 crc kubenswrapper[4856]: E0126 17:00:44.394692 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 17:00:45 crc kubenswrapper[4856]: I0126 17:00:45.395801 4856 scope.go:117] "RemoveContainer" containerID="afeb20035224feeab28a92ac77b43a24e653e49c56a25590a9861019a2b7a8ff" Jan 26 17:00:45 crc kubenswrapper[4856]: E0126 17:00:45.737235 4856 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Jan 26 17:00:46 crc kubenswrapper[4856]: I0126 17:00:46.053396 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-rq622_7a742e7b-c420-46e3-9e96-e9c744af6124/kube-multus/1.log" Jan 26 17:00:46 crc kubenswrapper[4856]: I0126 17:00:46.053613 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-rq622" event={"ID":"7a742e7b-c420-46e3-9e96-e9c744af6124","Type":"ContainerStarted","Data":"ddec0dbea657c6160cfdfd78886d5ae335dab8b667b0e0e3813dffa86a2ae2dc"} Jan 26 17:00:46 crc kubenswrapper[4856]: I0126 17:00:46.394810 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 17:00:46 crc kubenswrapper[4856]: I0126 17:00:46.394810 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 17:00:46 crc kubenswrapper[4856]: I0126 17:00:46.394818 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 17:00:46 crc kubenswrapper[4856]: I0126 17:00:46.394830 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 17:00:46 crc kubenswrapper[4856]: E0126 17:00:46.395209 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 17:00:46 crc kubenswrapper[4856]: E0126 17:00:46.395363 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 17:00:46 crc kubenswrapper[4856]: E0126 17:00:46.395566 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 17:00:46 crc kubenswrapper[4856]: E0126 17:00:46.395636 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 17:00:48 crc kubenswrapper[4856]: I0126 17:00:48.395118 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 17:00:48 crc kubenswrapper[4856]: I0126 17:00:48.395126 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 17:00:48 crc kubenswrapper[4856]: I0126 17:00:48.395183 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 17:00:48 crc kubenswrapper[4856]: E0126 17:00:48.396867 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 17:00:48 crc kubenswrapper[4856]: E0126 17:00:48.396991 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 17:00:48 crc kubenswrapper[4856]: I0126 17:00:48.395265 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 17:00:48 crc kubenswrapper[4856]: E0126 17:00:48.397245 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 17:00:48 crc kubenswrapper[4856]: E0126 17:00:48.397451 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 17:00:50 crc kubenswrapper[4856]: I0126 17:00:50.395394 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 17:00:50 crc kubenswrapper[4856]: I0126 17:00:50.395465 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 17:00:50 crc kubenswrapper[4856]: E0126 17:00:50.395596 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 17:00:50 crc kubenswrapper[4856]: I0126 17:00:50.395683 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 17:00:50 crc kubenswrapper[4856]: E0126 17:00:50.395718 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-295wr" podUID="12e50462-28e6-4531-ada4-e652310e6cce" Jan 26 17:00:50 crc kubenswrapper[4856]: E0126 17:00:50.395753 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 17:00:50 crc kubenswrapper[4856]: I0126 17:00:50.396004 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 17:00:50 crc kubenswrapper[4856]: E0126 17:00:50.396171 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.233908 4856 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.273535 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-6rlxp"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.274257 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.274257 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lndnt"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.275011 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-lndnt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.278670 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-6cghs"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.279258 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6cghs" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.281235 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.282511 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpqvc"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.283000 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpqvc" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.286786 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.287337 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.287387 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.287490 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.287580 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.287816 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.287923 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.287980 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.288106 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.288278 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.288334 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.288396 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.288399 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.288442 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.288419 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.288341 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.288565 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.288579 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.288596 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.288709 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.289021 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.289351 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.290359 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.290384 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.291517 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.291910 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4w5bf"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.292110 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.292591 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4w5bf" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.292909 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-7l927"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.293438 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-7l927" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.295482 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.295936 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.295955 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-5bjl7"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.296473 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-cb8nk"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.296774 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bjl7" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.298845 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.301024 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-6qgnn"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.301335 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7p5jt"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.301591 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.301724 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7p5jt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.301813 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-6qgnn" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.301629 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-962cr"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.302291 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-88lkr"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.302677 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-4pbj2"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.303032 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-4pbj2" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.303175 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-88lkr" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.303467 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-962cr" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.308809 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-jdjcq"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.309617 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-jdjcq" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.334071 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-7xb2b"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.336802 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-7xb2b" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.338175 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.338440 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.338667 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.339084 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.339698 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.339937 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.340183 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.340393 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.340798 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-27vjc"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.343443 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.343971 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.381408 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-27vjc" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.381930 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.381975 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.383035 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.384154 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.384417 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.384541 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.384708 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.384893 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.385846 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.386796 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.386854 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2ba3cf6a-a6be-4108-a155-c8bb530aa037-serving-cert\") pod \"openshift-config-operator-7777fb866f-5bjl7\" (UID: \"2ba3cf6a-a6be-4108-a155-c8bb530aa037\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bjl7" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.386915 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.386916 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/2ba3cf6a-a6be-4108-a155-c8bb530aa037-available-featuregates\") pod \"openshift-config-operator-7777fb866f-5bjl7\" (UID: \"2ba3cf6a-a6be-4108-a155-c8bb530aa037\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bjl7" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.386976 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tmd5\" (UniqueName: \"kubernetes.io/projected/2ba3cf6a-a6be-4108-a155-c8bb530aa037-kube-api-access-6tmd5\") pod \"openshift-config-operator-7777fb866f-5bjl7\" (UID: \"2ba3cf6a-a6be-4108-a155-c8bb530aa037\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bjl7" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.387096 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.387185 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.387206 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.387296 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.387404 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.387441 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.387492 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.387581 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.387226 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.390143 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.392517 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.392587 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.392724 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.392873 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.393005 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.393244 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.393860 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.394046 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.394181 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.394687 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.394851 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.395071 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.395152 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.392534 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.395470 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.395637 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.398315 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.398520 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.398914 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.399100 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.399143 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.399276 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.399315 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.400017 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.401564 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.406617 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.413414 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.414470 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.414660 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.414875 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.415175 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.415323 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.416204 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.416612 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-qdmxz"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.417136 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-58fcz"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.417313 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.417509 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.417545 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-q7j7b"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.417682 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.417835 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.417872 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-q7j7b" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.418099 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qdmxz" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.418264 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-58fcz" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.437074 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.437937 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.439191 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.439792 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.440562 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.444734 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.447356 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-h9b2g"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.447709 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.447890 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.448346 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-h9b2g" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.449829 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6snv6"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.450230 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6snv6" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.456403 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.459349 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cl895"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.460042 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cl895" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.461277 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-2sfhr"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.461987 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2sfhr" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.462128 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-zzxln"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.463902 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-ddghz"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.464380 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-ddghz" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.464420 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-zzxln" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.465694 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k662z"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.466487 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k662z" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.466502 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nn46h"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.467185 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nn46h" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.468450 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8m4l6"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.469276 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8m4l6" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.470562 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mr7cp"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.470916 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mr7cp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.472281 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lndnt"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.473233 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpqvc"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.475507 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-gz7kg"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.475832 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l9nqd"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.476221 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l9nqd" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.476741 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-gz7kg" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.479609 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wvttb"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.480363 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wvttb" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.480915 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490780-8q6q4"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.481684 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-8q6q4" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.483168 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rrhjv"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.483768 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rrhjv" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.484671 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-cjzsq"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.485338 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-cjzsq" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.486437 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vmwvg"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.486943 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vmwvg" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.488017 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-wxbdh"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.488149 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lg98g\" (UniqueName: \"kubernetes.io/projected/a6d331bd-2db3-4319-9f5c-db56d408d9e3-kube-api-access-lg98g\") pod \"apiserver-76f77b778f-6rlxp\" (UID: \"a6d331bd-2db3-4319-9f5c-db56d408d9e3\") " pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.488196 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/81c2f96b-55e0-483b-b72c-df7e156e9218-audit-policies\") pod \"apiserver-7bbb656c7d-6cghs\" (UID: \"81c2f96b-55e0-483b-b72c-df7e156e9218\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6cghs" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.488446 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.488515 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a7e5d16a-d45d-4b40-93bf-bfaa6be2d1c0-service-ca-bundle\") pod \"authentication-operator-69f744f599-jdjcq\" (UID: \"a7e5d16a-d45d-4b40-93bf-bfaa6be2d1c0\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jdjcq" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.488633 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54a246a2-f674-4735-b295-b56699ece95b-config\") pod \"machine-approver-56656f9798-962cr\" (UID: \"54a246a2-f674-4735-b295-b56699ece95b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-962cr" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.488755 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mr8gn\" (UniqueName: \"kubernetes.io/projected/1afc0f4c-e02d-4a70-aaba-e761e8c04eee-kube-api-access-mr8gn\") pod \"controller-manager-879f6c89f-lndnt\" (UID: \"1afc0f4c-e02d-4a70-aaba-e761e8c04eee\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lndnt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.489354 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/42c0e428-821f-45a1-85a7-54ebdb81ef1c-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cl895\" (UID: \"42c0e428-821f-45a1-85a7-54ebdb81ef1c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cl895" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.489386 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa11789e-7a2a-4dbf-85ca-c20a9d64a1f4-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-6snv6\" (UID: \"fa11789e-7a2a-4dbf-85ca-c20a9d64a1f4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6snv6" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.489401 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf1f11c8-17b8-49b7-b12d-92891f478222-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-7p5jt\" (UID: \"bf1f11c8-17b8-49b7-b12d-92891f478222\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7p5jt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.489417 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpfwk\" (UniqueName: \"kubernetes.io/projected/5fe6baed-ab97-4d8a-8be2-6f00f9698136-kube-api-access-hpfwk\") pod \"route-controller-manager-6576b87f9c-fpqvc\" (UID: \"5fe6baed-ab97-4d8a-8be2-6f00f9698136\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpqvc" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.489437 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5c244eff-aada-44f3-b250-96878a3400c4-etcd-client\") pod \"etcd-operator-b45778765-27vjc\" (UID: \"5c244eff-aada-44f3-b250-96878a3400c4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-27vjc" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.489454 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf1f11c8-17b8-49b7-b12d-92891f478222-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-7p5jt\" (UID: \"bf1f11c8-17b8-49b7-b12d-92891f478222\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7p5jt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.489472 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-cb8nk\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.489489 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b28404ed-2e71-4b3f-9140-35ee89dbc8f2-console-config\") pod \"console-f9d7485db-6qgnn\" (UID: \"b28404ed-2e71-4b3f-9140-35ee89dbc8f2\") " pod="openshift-console/console-f9d7485db-6qgnn" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.489514 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a6d331bd-2db3-4319-9f5c-db56d408d9e3-serving-cert\") pod \"apiserver-76f77b778f-6rlxp\" (UID: \"a6d331bd-2db3-4319-9f5c-db56d408d9e3\") " pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.489551 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-cb8nk\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.489566 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-cb8nk\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.489655 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a7e5d16a-d45d-4b40-93bf-bfaa6be2d1c0-serving-cert\") pod \"authentication-operator-69f744f599-jdjcq\" (UID: \"a7e5d16a-d45d-4b40-93bf-bfaa6be2d1c0\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jdjcq" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.489693 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fe6baed-ab97-4d8a-8be2-6f00f9698136-config\") pod \"route-controller-manager-6576b87f9c-fpqvc\" (UID: \"5fe6baed-ab97-4d8a-8be2-6f00f9698136\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpqvc" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.489723 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-cb8nk\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.489821 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b28404ed-2e71-4b3f-9140-35ee89dbc8f2-oauth-serving-cert\") pod \"console-f9d7485db-6qgnn\" (UID: \"b28404ed-2e71-4b3f-9140-35ee89dbc8f2\") " pod="openshift-console/console-f9d7485db-6qgnn" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.490070 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42c0e428-821f-45a1-85a7-54ebdb81ef1c-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cl895\" (UID: \"42c0e428-821f-45a1-85a7-54ebdb81ef1c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cl895" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.490098 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqhmb\" (UniqueName: \"kubernetes.io/projected/81c2f96b-55e0-483b-b72c-df7e156e9218-kube-api-access-rqhmb\") pod \"apiserver-7bbb656c7d-6cghs\" (UID: \"81c2f96b-55e0-483b-b72c-df7e156e9218\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6cghs" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.490119 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/149e3000-35d7-47bd-83f0-00ab5e0736c2-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-q7j7b\" (UID: \"149e3000-35d7-47bd-83f0-00ab5e0736c2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-q7j7b" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.490235 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa11789e-7a2a-4dbf-85ca-c20a9d64a1f4-config\") pod \"kube-controller-manager-operator-78b949d7b-6snv6\" (UID: \"fa11789e-7a2a-4dbf-85ca-c20a9d64a1f4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6snv6" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.490282 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/2ba3cf6a-a6be-4108-a155-c8bb530aa037-available-featuregates\") pod \"openshift-config-operator-7777fb866f-5bjl7\" (UID: \"2ba3cf6a-a6be-4108-a155-c8bb530aa037\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bjl7" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.490308 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/a6d331bd-2db3-4319-9f5c-db56d408d9e3-image-import-ca\") pod \"apiserver-76f77b778f-6rlxp\" (UID: \"a6d331bd-2db3-4319-9f5c-db56d408d9e3\") " pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.490343 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/81c2f96b-55e0-483b-b72c-df7e156e9218-audit-dir\") pod \"apiserver-7bbb656c7d-6cghs\" (UID: \"81c2f96b-55e0-483b-b72c-df7e156e9218\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6cghs" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.490373 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7e5d16a-d45d-4b40-93bf-bfaa6be2d1c0-config\") pod \"authentication-operator-69f744f599-jdjcq\" (UID: \"a7e5d16a-d45d-4b40-93bf-bfaa6be2d1c0\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jdjcq" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.490397 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/831dc87e-8e14-43d3-a36e-dc7679041ae5-config\") pod \"console-operator-58897d9998-4pbj2\" (UID: \"831dc87e-8e14-43d3-a36e-dc7679041ae5\") " pod="openshift-console-operator/console-operator-58897d9998-4pbj2" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.490430 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b28404ed-2e71-4b3f-9140-35ee89dbc8f2-trusted-ca-bundle\") pod \"console-f9d7485db-6qgnn\" (UID: \"b28404ed-2e71-4b3f-9140-35ee89dbc8f2\") " pod="openshift-console/console-f9d7485db-6qgnn" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.490468 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkl7p\" (UniqueName: \"kubernetes.io/projected/359660cd-b412-4640-bedf-993e976e7b3c-kube-api-access-rkl7p\") pod \"openshift-apiserver-operator-796bbdcf4f-88lkr\" (UID: \"359660cd-b412-4640-bedf-993e976e7b3c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-88lkr" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.490499 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qcw5\" (UniqueName: \"kubernetes.io/projected/bf1f11c8-17b8-49b7-b12d-92891f478222-kube-api-access-9qcw5\") pod \"cluster-image-registry-operator-dc59b4c8b-7p5jt\" (UID: \"bf1f11c8-17b8-49b7-b12d-92891f478222\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7p5jt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.490541 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2wtm\" (UniqueName: \"kubernetes.io/projected/0c1af7db-aa80-4cb0-a9cb-5afdf677f28c-kube-api-access-v2wtm\") pod \"cluster-samples-operator-665b6dd947-4w5bf\" (UID: \"0c1af7db-aa80-4cb0-a9cb-5afdf677f28c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4w5bf" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.490568 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkngl\" (UniqueName: \"kubernetes.io/projected/5c244eff-aada-44f3-b250-96878a3400c4-kube-api-access-nkngl\") pod \"etcd-operator-b45778765-27vjc\" (UID: \"5c244eff-aada-44f3-b250-96878a3400c4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-27vjc" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.490593 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/69008ed1-f3e5-400d-852f-adbcd94199f6-audit-dir\") pod \"oauth-openshift-558db77b4-cb8nk\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.490612 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1afc0f4c-e02d-4a70-aaba-e761e8c04eee-client-ca\") pod \"controller-manager-879f6c89f-lndnt\" (UID: \"1afc0f4c-e02d-4a70-aaba-e761e8c04eee\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lndnt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.490646 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f9b8f57e-00b9-4355-ace2-0319d320d208-webhook-cert\") pod \"packageserver-d55dfcdfc-mr7cp\" (UID: \"f9b8f57e-00b9-4355-ace2-0319d320d208\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mr7cp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.490654 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/2ba3cf6a-a6be-4108-a155-c8bb530aa037-available-featuregates\") pod \"openshift-config-operator-7777fb866f-5bjl7\" (UID: \"2ba3cf6a-a6be-4108-a155-c8bb530aa037\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bjl7" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.490674 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vq4z\" (UniqueName: \"kubernetes.io/projected/a7e5d16a-d45d-4b40-93bf-bfaa6be2d1c0-kube-api-access-7vq4z\") pod \"authentication-operator-69f744f599-jdjcq\" (UID: \"a7e5d16a-d45d-4b40-93bf-bfaa6be2d1c0\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jdjcq" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.490707 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/0c1af7db-aa80-4cb0-a9cb-5afdf677f28c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-4w5bf\" (UID: \"0c1af7db-aa80-4cb0-a9cb-5afdf677f28c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4w5bf" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.490730 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/42c0e428-821f-45a1-85a7-54ebdb81ef1c-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cl895\" (UID: \"42c0e428-821f-45a1-85a7-54ebdb81ef1c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cl895" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.490763 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgb6k\" (UniqueName: \"kubernetes.io/projected/94291fa4-24a5-499e-8143-89c8784d9284-kube-api-access-hgb6k\") pod \"downloads-7954f5f757-7l927\" (UID: \"94291fa4-24a5-499e-8143-89c8784d9284\") " pod="openshift-console/downloads-7954f5f757-7l927" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.490785 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ddc2e6b7-5582-4579-bf2c-ed165b74c91a-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-vmwvg\" (UID: \"ddc2e6b7-5582-4579-bf2c-ed165b74c91a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vmwvg" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.490818 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/81c2f96b-55e0-483b-b72c-df7e156e9218-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-6cghs\" (UID: \"81c2f96b-55e0-483b-b72c-df7e156e9218\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6cghs" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.490834 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjhr5\" (UniqueName: \"kubernetes.io/projected/beb6f283-75cb-4184-b985-4e6c095feca1-kube-api-access-mjhr5\") pod \"multus-admission-controller-857f4d67dd-ddghz\" (UID: \"beb6f283-75cb-4184-b985-4e6c095feca1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ddghz" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.490850 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05d74105-0ecd-41ac-9001-8b21b0fd6ba4-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-l9nqd\" (UID: \"05d74105-0ecd-41ac-9001-8b21b0fd6ba4\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l9nqd" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.490865 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/85f05bd5-ff83-4d29-9531-ab3499088095-metrics-certs\") pod \"router-default-5444994796-h9b2g\" (UID: \"85f05bd5-ff83-4d29-9531-ab3499088095\") " pod="openshift-ingress/router-default-5444994796-h9b2g" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.490882 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7lk4\" (UniqueName: \"kubernetes.io/projected/85f05bd5-ff83-4d29-9531-ab3499088095-kube-api-access-x7lk4\") pod \"router-default-5444994796-h9b2g\" (UID: \"85f05bd5-ff83-4d29-9531-ab3499088095\") " pod="openshift-ingress/router-default-5444994796-h9b2g" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.490897 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/149e3000-35d7-47bd-83f0-00ab5e0736c2-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-q7j7b\" (UID: \"149e3000-35d7-47bd-83f0-00ab5e0736c2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-q7j7b" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.490913 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6d331bd-2db3-4319-9f5c-db56d408d9e3-config\") pod \"apiserver-76f77b778f-6rlxp\" (UID: \"a6d331bd-2db3-4319-9f5c-db56d408d9e3\") " pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.490928 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ddc2e6b7-5582-4579-bf2c-ed165b74c91a-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-vmwvg\" (UID: \"ddc2e6b7-5582-4579-bf2c-ed165b74c91a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vmwvg" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.490946 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tmd5\" (UniqueName: \"kubernetes.io/projected/2ba3cf6a-a6be-4108-a155-c8bb530aa037-kube-api-access-6tmd5\") pod \"openshift-config-operator-7777fb866f-5bjl7\" (UID: \"2ba3cf6a-a6be-4108-a155-c8bb530aa037\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bjl7" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.490961 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a6d331bd-2db3-4319-9f5c-db56d408d9e3-etcd-client\") pod \"apiserver-76f77b778f-6rlxp\" (UID: \"a6d331bd-2db3-4319-9f5c-db56d408d9e3\") " pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.490979 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/54a246a2-f674-4735-b295-b56699ece95b-machine-approver-tls\") pod \"machine-approver-56656f9798-962cr\" (UID: \"54a246a2-f674-4735-b295-b56699ece95b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-962cr" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.491001 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-cb8nk\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.491017 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/54a246a2-f674-4735-b295-b56699ece95b-auth-proxy-config\") pod \"machine-approver-56656f9798-962cr\" (UID: \"54a246a2-f674-4735-b295-b56699ece95b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-962cr" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.491053 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6fxv\" (UniqueName: \"kubernetes.io/projected/f9b8f57e-00b9-4355-ace2-0319d320d208-kube-api-access-d6fxv\") pod \"packageserver-d55dfcdfc-mr7cp\" (UID: \"f9b8f57e-00b9-4355-ace2-0319d320d208\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mr7cp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.491069 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/831dc87e-8e14-43d3-a36e-dc7679041ae5-trusted-ca\") pod \"console-operator-58897d9998-4pbj2\" (UID: \"831dc87e-8e14-43d3-a36e-dc7679041ae5\") " pod="openshift-console-operator/console-operator-58897d9998-4pbj2" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.491082 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/f9b8f57e-00b9-4355-ace2-0319d320d208-tmpfs\") pod \"packageserver-d55dfcdfc-mr7cp\" (UID: \"f9b8f57e-00b9-4355-ace2-0319d320d208\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mr7cp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.491102 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/033cb12f-278f-431a-8104-519db9a3152f-signing-key\") pod \"service-ca-9c57cc56f-gz7kg\" (UID: \"033cb12f-278f-431a-8104-519db9a3152f\") " pod="openshift-service-ca/service-ca-9c57cc56f-gz7kg" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.491117 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/5c244eff-aada-44f3-b250-96878a3400c4-etcd-service-ca\") pod \"etcd-operator-b45778765-27vjc\" (UID: \"5c244eff-aada-44f3-b250-96878a3400c4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-27vjc" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.491155 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a7e5d16a-d45d-4b40-93bf-bfaa6be2d1c0-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-jdjcq\" (UID: \"a7e5d16a-d45d-4b40-93bf-bfaa6be2d1c0\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jdjcq" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.491169 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2274g\" (UniqueName: \"kubernetes.io/projected/54a246a2-f674-4735-b295-b56699ece95b-kube-api-access-2274g\") pod \"machine-approver-56656f9798-962cr\" (UID: \"54a246a2-f674-4735-b295-b56699ece95b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-962cr" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.491364 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/85f05bd5-ff83-4d29-9531-ab3499088095-stats-auth\") pod \"router-default-5444994796-h9b2g\" (UID: \"85f05bd5-ff83-4d29-9531-ab3499088095\") " pod="openshift-ingress/router-default-5444994796-h9b2g" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.491380 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2wnd\" (UniqueName: \"kubernetes.io/projected/831dc87e-8e14-43d3-a36e-dc7679041ae5-kube-api-access-d2wnd\") pod \"console-operator-58897d9998-4pbj2\" (UID: \"831dc87e-8e14-43d3-a36e-dc7679041ae5\") " pod="openshift-console-operator/console-operator-58897d9998-4pbj2" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.491395 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77a97acb-2908-48fb-8bcd-0647f3e90160-config\") pod \"machine-api-operator-5694c8668f-7xb2b\" (UID: \"77a97acb-2908-48fb-8bcd-0647f3e90160\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7xb2b" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.491409 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c475g\" (UniqueName: \"kubernetes.io/projected/77a97acb-2908-48fb-8bcd-0647f3e90160-kube-api-access-c475g\") pod \"machine-api-operator-5694c8668f-7xb2b\" (UID: \"77a97acb-2908-48fb-8bcd-0647f3e90160\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7xb2b" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.491425 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-cb8nk\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.491438 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b28404ed-2e71-4b3f-9140-35ee89dbc8f2-console-oauth-config\") pod \"console-f9d7485db-6qgnn\" (UID: \"b28404ed-2e71-4b3f-9140-35ee89dbc8f2\") " pod="openshift-console/console-f9d7485db-6qgnn" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.491452 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b28404ed-2e71-4b3f-9140-35ee89dbc8f2-service-ca\") pod \"console-f9d7485db-6qgnn\" (UID: \"b28404ed-2e71-4b3f-9140-35ee89dbc8f2\") " pod="openshift-console/console-f9d7485db-6qgnn" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.491464 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f9b8f57e-00b9-4355-ace2-0319d320d208-apiservice-cert\") pod \"packageserver-d55dfcdfc-mr7cp\" (UID: \"f9b8f57e-00b9-4355-ace2-0319d320d208\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mr7cp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.491478 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/beb6f283-75cb-4184-b985-4e6c095feca1-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-ddghz\" (UID: \"beb6f283-75cb-4184-b985-4e6c095feca1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ddghz" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.491491 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/69008ed1-f3e5-400d-852f-adbcd94199f6-audit-policies\") pod \"oauth-openshift-558db77b4-cb8nk\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.491505 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-cb8nk\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.491518 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b28404ed-2e71-4b3f-9140-35ee89dbc8f2-console-serving-cert\") pod \"console-f9d7485db-6qgnn\" (UID: \"b28404ed-2e71-4b3f-9140-35ee89dbc8f2\") " pod="openshift-console/console-f9d7485db-6qgnn" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.491568 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/77a97acb-2908-48fb-8bcd-0647f3e90160-images\") pod \"machine-api-operator-5694c8668f-7xb2b\" (UID: \"77a97acb-2908-48fb-8bcd-0647f3e90160\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7xb2b" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.491585 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/77a97acb-2908-48fb-8bcd-0647f3e90160-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-7xb2b\" (UID: \"77a97acb-2908-48fb-8bcd-0647f3e90160\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7xb2b" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.491600 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-cb8nk\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.491616 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/5c244eff-aada-44f3-b250-96878a3400c4-etcd-ca\") pod \"etcd-operator-b45778765-27vjc\" (UID: \"5c244eff-aada-44f3-b250-96878a3400c4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-27vjc" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.491630 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/359660cd-b412-4640-bedf-993e976e7b3c-config\") pod \"openshift-apiserver-operator-796bbdcf4f-88lkr\" (UID: \"359660cd-b412-4640-bedf-993e976e7b3c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-88lkr" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.491646 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddc2e6b7-5582-4579-bf2c-ed165b74c91a-config\") pod \"kube-apiserver-operator-766d6c64bb-vmwvg\" (UID: \"ddc2e6b7-5582-4579-bf2c-ed165b74c91a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vmwvg" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.491662 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2ba3cf6a-a6be-4108-a155-c8bb530aa037-serving-cert\") pod \"openshift-config-operator-7777fb866f-5bjl7\" (UID: \"2ba3cf6a-a6be-4108-a155-c8bb530aa037\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bjl7" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.491677 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c244eff-aada-44f3-b250-96878a3400c4-serving-cert\") pod \"etcd-operator-b45778765-27vjc\" (UID: \"5c244eff-aada-44f3-b250-96878a3400c4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-27vjc" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.491693 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/81c2f96b-55e0-483b-b72c-df7e156e9218-serving-cert\") pod \"apiserver-7bbb656c7d-6cghs\" (UID: \"81c2f96b-55e0-483b-b72c-df7e156e9218\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6cghs" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.491711 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-cb8nk\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.491733 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c244eff-aada-44f3-b250-96878a3400c4-config\") pod \"etcd-operator-b45778765-27vjc\" (UID: \"5c244eff-aada-44f3-b250-96878a3400c4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-27vjc" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.491753 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/81c2f96b-55e0-483b-b72c-df7e156e9218-encryption-config\") pod \"apiserver-7bbb656c7d-6cghs\" (UID: \"81c2f96b-55e0-483b-b72c-df7e156e9218\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6cghs" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.491773 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-cb8nk\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.491794 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fa11789e-7a2a-4dbf-85ca-c20a9d64a1f4-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-6snv6\" (UID: \"fa11789e-7a2a-4dbf-85ca-c20a9d64a1f4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6snv6" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.491815 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/a6d331bd-2db3-4319-9f5c-db56d408d9e3-encryption-config\") pod \"apiserver-76f77b778f-6rlxp\" (UID: \"a6d331bd-2db3-4319-9f5c-db56d408d9e3\") " pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.491834 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf1f11c8-17b8-49b7-b12d-92891f478222-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-7p5jt\" (UID: \"bf1f11c8-17b8-49b7-b12d-92891f478222\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7p5jt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.491866 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kf2b2\" (UniqueName: \"kubernetes.io/projected/69008ed1-f3e5-400d-852f-adbcd94199f6-kube-api-access-kf2b2\") pod \"oauth-openshift-558db77b4-cb8nk\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.491900 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1afc0f4c-e02d-4a70-aaba-e761e8c04eee-serving-cert\") pod \"controller-manager-879f6c89f-lndnt\" (UID: \"1afc0f4c-e02d-4a70-aaba-e761e8c04eee\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lndnt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.491922 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a6d331bd-2db3-4319-9f5c-db56d408d9e3-node-pullsecrets\") pod \"apiserver-76f77b778f-6rlxp\" (UID: \"a6d331bd-2db3-4319-9f5c-db56d408d9e3\") " pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.491941 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/a6d331bd-2db3-4319-9f5c-db56d408d9e3-audit\") pod \"apiserver-76f77b778f-6rlxp\" (UID: \"a6d331bd-2db3-4319-9f5c-db56d408d9e3\") " pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.491961 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/81c2f96b-55e0-483b-b72c-df7e156e9218-etcd-client\") pod \"apiserver-7bbb656c7d-6cghs\" (UID: \"81c2f96b-55e0-483b-b72c-df7e156e9218\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6cghs" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.491982 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/831dc87e-8e14-43d3-a36e-dc7679041ae5-serving-cert\") pod \"console-operator-58897d9998-4pbj2\" (UID: \"831dc87e-8e14-43d3-a36e-dc7679041ae5\") " pod="openshift-console-operator/console-operator-58897d9998-4pbj2" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.492003 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/85f05bd5-ff83-4d29-9531-ab3499088095-default-certificate\") pod \"router-default-5444994796-h9b2g\" (UID: \"85f05bd5-ff83-4d29-9531-ab3499088095\") " pod="openshift-ingress/router-default-5444994796-h9b2g" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.492026 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/a6d331bd-2db3-4319-9f5c-db56d408d9e3-etcd-serving-ca\") pod \"apiserver-76f77b778f-6rlxp\" (UID: \"a6d331bd-2db3-4319-9f5c-db56d408d9e3\") " pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.492046 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5fe6baed-ab97-4d8a-8be2-6f00f9698136-serving-cert\") pod \"route-controller-manager-6576b87f9c-fpqvc\" (UID: \"5fe6baed-ab97-4d8a-8be2-6f00f9698136\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpqvc" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.492068 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-cb8nk\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.492092 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5fe6baed-ab97-4d8a-8be2-6f00f9698136-client-ca\") pod \"route-controller-manager-6576b87f9c-fpqvc\" (UID: \"5fe6baed-ab97-4d8a-8be2-6f00f9698136\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpqvc" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.492112 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1afc0f4c-e02d-4a70-aaba-e761e8c04eee-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-lndnt\" (UID: \"1afc0f4c-e02d-4a70-aaba-e761e8c04eee\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lndnt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.492134 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvkhn\" (UniqueName: \"kubernetes.io/projected/b28404ed-2e71-4b3f-9140-35ee89dbc8f2-kube-api-access-dvkhn\") pod \"console-f9d7485db-6qgnn\" (UID: \"b28404ed-2e71-4b3f-9140-35ee89dbc8f2\") " pod="openshift-console/console-f9d7485db-6qgnn" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.492155 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/033cb12f-278f-431a-8104-519db9a3152f-signing-cabundle\") pod \"service-ca-9c57cc56f-gz7kg\" (UID: \"033cb12f-278f-431a-8104-519db9a3152f\") " pod="openshift-service-ca/service-ca-9c57cc56f-gz7kg" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.492175 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a6d331bd-2db3-4319-9f5c-db56d408d9e3-trusted-ca-bundle\") pod \"apiserver-76f77b778f-6rlxp\" (UID: \"a6d331bd-2db3-4319-9f5c-db56d408d9e3\") " pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.492196 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5scg\" (UniqueName: \"kubernetes.io/projected/033cb12f-278f-431a-8104-519db9a3152f-kube-api-access-s5scg\") pod \"service-ca-9c57cc56f-gz7kg\" (UID: \"033cb12f-278f-431a-8104-519db9a3152f\") " pod="openshift-service-ca/service-ca-9c57cc56f-gz7kg" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.492214 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdh5l\" (UniqueName: \"kubernetes.io/projected/149e3000-35d7-47bd-83f0-00ab5e0736c2-kube-api-access-mdh5l\") pod \"kube-storage-version-migrator-operator-b67b599dd-q7j7b\" (UID: \"149e3000-35d7-47bd-83f0-00ab5e0736c2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-q7j7b" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.492232 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/85f05bd5-ff83-4d29-9531-ab3499088095-service-ca-bundle\") pod \"router-default-5444994796-h9b2g\" (UID: \"85f05bd5-ff83-4d29-9531-ab3499088095\") " pod="openshift-ingress/router-default-5444994796-h9b2g" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.492251 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05d74105-0ecd-41ac-9001-8b21b0fd6ba4-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-l9nqd\" (UID: \"05d74105-0ecd-41ac-9001-8b21b0fd6ba4\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l9nqd" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.492272 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a6d331bd-2db3-4319-9f5c-db56d408d9e3-audit-dir\") pod \"apiserver-76f77b778f-6rlxp\" (UID: \"a6d331bd-2db3-4319-9f5c-db56d408d9e3\") " pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.492291 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/81c2f96b-55e0-483b-b72c-df7e156e9218-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-6cghs\" (UID: \"81c2f96b-55e0-483b-b72c-df7e156e9218\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6cghs" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.492312 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/359660cd-b412-4640-bedf-993e976e7b3c-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-88lkr\" (UID: \"359660cd-b412-4640-bedf-993e976e7b3c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-88lkr" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.492331 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1afc0f4c-e02d-4a70-aaba-e761e8c04eee-config\") pod \"controller-manager-879f6c89f-lndnt\" (UID: \"1afc0f4c-e02d-4a70-aaba-e761e8c04eee\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lndnt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.494960 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-7l927"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.496953 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.497120 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-z7cgq"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.497902 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-z7cgq" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.499997 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-fbsj7"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.500407 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-fbsj7" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.501050 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2ba3cf6a-a6be-4108-a155-c8bb530aa037-serving-cert\") pod \"openshift-config-operator-7777fb866f-5bjl7\" (UID: \"2ba3cf6a-a6be-4108-a155-c8bb530aa037\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bjl7" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.502055 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-cb8nk"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.503833 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-6rlxp"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.505611 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7p5jt"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.508574 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-58fcz"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.509783 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4w5bf"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.511820 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-zzxln"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.514340 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cl895"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.515956 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-4pbj2"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.519450 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.519869 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-2sfhr"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.521904 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-5bjl7"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.522808 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mr7cp"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.524060 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-27vjc"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.525403 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-7xb2b"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.526786 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-jdjcq"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.534207 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k662z"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.536496 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-ddghz"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.537383 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.538024 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-qdmxz"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.539666 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-6qgnn"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.541667 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-6cghs"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.543466 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nn46h"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.544777 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vmwvg"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.546517 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-wxbdh"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.548399 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8m4l6"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.550491 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6snv6"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.551425 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-88lkr"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.553235 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-gz7kg"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.554615 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l9nqd"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.556805 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.556872 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-q7j7b"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.557677 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rrhjv"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.559074 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-vfm8t"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.560431 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-vfm8t" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.560769 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-c9qdp"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.561294 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-c9qdp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.561956 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-fbsj7"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.563301 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-cjzsq"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.564760 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-z7cgq"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.566190 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wvttb"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.567598 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490780-8q6q4"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.569111 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-vfm8t"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.570422 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-dgcqn"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.571167 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-dgcqn" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.571566 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-dgcqn"] Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.575619 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.593558 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdh5l\" (UniqueName: \"kubernetes.io/projected/149e3000-35d7-47bd-83f0-00ab5e0736c2-kube-api-access-mdh5l\") pod \"kube-storage-version-migrator-operator-b67b599dd-q7j7b\" (UID: \"149e3000-35d7-47bd-83f0-00ab5e0736c2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-q7j7b" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.593605 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1afc0f4c-e02d-4a70-aaba-e761e8c04eee-config\") pod \"controller-manager-879f6c89f-lndnt\" (UID: \"1afc0f4c-e02d-4a70-aaba-e761e8c04eee\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lndnt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.593630 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/85f05bd5-ff83-4d29-9531-ab3499088095-service-ca-bundle\") pod \"router-default-5444994796-h9b2g\" (UID: \"85f05bd5-ff83-4d29-9531-ab3499088095\") " pod="openshift-ingress/router-default-5444994796-h9b2g" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.593652 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a6d331bd-2db3-4319-9f5c-db56d408d9e3-audit-dir\") pod \"apiserver-76f77b778f-6rlxp\" (UID: \"a6d331bd-2db3-4319-9f5c-db56d408d9e3\") " pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.593676 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/81c2f96b-55e0-483b-b72c-df7e156e9218-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-6cghs\" (UID: \"81c2f96b-55e0-483b-b72c-df7e156e9218\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6cghs" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.593696 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/81c2f96b-55e0-483b-b72c-df7e156e9218-audit-policies\") pod \"apiserver-7bbb656c7d-6cghs\" (UID: \"81c2f96b-55e0-483b-b72c-df7e156e9218\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6cghs" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.593712 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a7e5d16a-d45d-4b40-93bf-bfaa6be2d1c0-service-ca-bundle\") pod \"authentication-operator-69f744f599-jdjcq\" (UID: \"a7e5d16a-d45d-4b40-93bf-bfaa6be2d1c0\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jdjcq" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.593813 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a6d331bd-2db3-4319-9f5c-db56d408d9e3-audit-dir\") pod \"apiserver-76f77b778f-6rlxp\" (UID: \"a6d331bd-2db3-4319-9f5c-db56d408d9e3\") " pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.594594 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/81c2f96b-55e0-483b-b72c-df7e156e9218-audit-policies\") pod \"apiserver-7bbb656c7d-6cghs\" (UID: \"81c2f96b-55e0-483b-b72c-df7e156e9218\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6cghs" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.594663 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54a246a2-f674-4735-b295-b56699ece95b-config\") pod \"machine-approver-56656f9798-962cr\" (UID: \"54a246a2-f674-4735-b295-b56699ece95b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-962cr" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.594684 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mr8gn\" (UniqueName: \"kubernetes.io/projected/1afc0f4c-e02d-4a70-aaba-e761e8c04eee-kube-api-access-mr8gn\") pod \"controller-manager-879f6c89f-lndnt\" (UID: \"1afc0f4c-e02d-4a70-aaba-e761e8c04eee\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lndnt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.594693 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/81c2f96b-55e0-483b-b72c-df7e156e9218-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-6cghs\" (UID: \"81c2f96b-55e0-483b-b72c-df7e156e9218\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6cghs" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.594775 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a7e5d16a-d45d-4b40-93bf-bfaa6be2d1c0-service-ca-bundle\") pod \"authentication-operator-69f744f599-jdjcq\" (UID: \"a7e5d16a-d45d-4b40-93bf-bfaa6be2d1c0\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jdjcq" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.595090 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54a246a2-f674-4735-b295-b56699ece95b-config\") pod \"machine-approver-56656f9798-962cr\" (UID: \"54a246a2-f674-4735-b295-b56699ece95b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-962cr" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.594700 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/42c0e428-821f-45a1-85a7-54ebdb81ef1c-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cl895\" (UID: \"42c0e428-821f-45a1-85a7-54ebdb81ef1c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cl895" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.595189 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lg98g\" (UniqueName: \"kubernetes.io/projected/a6d331bd-2db3-4319-9f5c-db56d408d9e3-kube-api-access-lg98g\") pod \"apiserver-76f77b778f-6rlxp\" (UID: \"a6d331bd-2db3-4319-9f5c-db56d408d9e3\") " pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.595206 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa11789e-7a2a-4dbf-85ca-c20a9d64a1f4-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-6snv6\" (UID: \"fa11789e-7a2a-4dbf-85ca-c20a9d64a1f4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6snv6" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.595225 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf1f11c8-17b8-49b7-b12d-92891f478222-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-7p5jt\" (UID: \"bf1f11c8-17b8-49b7-b12d-92891f478222\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7p5jt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.595324 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1afc0f4c-e02d-4a70-aaba-e761e8c04eee-config\") pod \"controller-manager-879f6c89f-lndnt\" (UID: \"1afc0f4c-e02d-4a70-aaba-e761e8c04eee\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lndnt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.595379 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpfwk\" (UniqueName: \"kubernetes.io/projected/5fe6baed-ab97-4d8a-8be2-6f00f9698136-kube-api-access-hpfwk\") pod \"route-controller-manager-6576b87f9c-fpqvc\" (UID: \"5fe6baed-ab97-4d8a-8be2-6f00f9698136\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpqvc" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.595399 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf1f11c8-17b8-49b7-b12d-92891f478222-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-7p5jt\" (UID: \"bf1f11c8-17b8-49b7-b12d-92891f478222\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7p5jt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.595522 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.595632 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a6d331bd-2db3-4319-9f5c-db56d408d9e3-serving-cert\") pod \"apiserver-76f77b778f-6rlxp\" (UID: \"a6d331bd-2db3-4319-9f5c-db56d408d9e3\") " pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.595659 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-cb8nk\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.595682 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/abbeffe1-cfd5-4476-9a8e-2ab5b4869444-profile-collector-cert\") pod \"catalog-operator-68c6474976-nn46h\" (UID: \"abbeffe1-cfd5-4476-9a8e-2ab5b4869444\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nn46h" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.595699 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a7e5d16a-d45d-4b40-93bf-bfaa6be2d1c0-serving-cert\") pod \"authentication-operator-69f744f599-jdjcq\" (UID: \"a7e5d16a-d45d-4b40-93bf-bfaa6be2d1c0\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jdjcq" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.595715 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fe6baed-ab97-4d8a-8be2-6f00f9698136-config\") pod \"route-controller-manager-6576b87f9c-fpqvc\" (UID: \"5fe6baed-ab97-4d8a-8be2-6f00f9698136\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpqvc" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.595740 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b28404ed-2e71-4b3f-9140-35ee89dbc8f2-oauth-serving-cert\") pod \"console-f9d7485db-6qgnn\" (UID: \"b28404ed-2e71-4b3f-9140-35ee89dbc8f2\") " pod="openshift-console/console-f9d7485db-6qgnn" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.595760 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wxzv\" (UniqueName: \"kubernetes.io/projected/17a72e73-4d54-4a29-a85a-ecb1aff30d10-kube-api-access-9wxzv\") pod \"olm-operator-6b444d44fb-k662z\" (UID: \"17a72e73-4d54-4a29-a85a-ecb1aff30d10\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k662z" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.595779 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa11789e-7a2a-4dbf-85ca-c20a9d64a1f4-config\") pod \"kube-controller-manager-operator-78b949d7b-6snv6\" (UID: \"fa11789e-7a2a-4dbf-85ca-c20a9d64a1f4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6snv6" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.595796 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/831dc87e-8e14-43d3-a36e-dc7679041ae5-config\") pod \"console-operator-58897d9998-4pbj2\" (UID: \"831dc87e-8e14-43d3-a36e-dc7679041ae5\") " pod="openshift-console-operator/console-operator-58897d9998-4pbj2" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.595810 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b28404ed-2e71-4b3f-9140-35ee89dbc8f2-trusted-ca-bundle\") pod \"console-f9d7485db-6qgnn\" (UID: \"b28404ed-2e71-4b3f-9140-35ee89dbc8f2\") " pod="openshift-console/console-f9d7485db-6qgnn" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.596084 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7e5d16a-d45d-4b40-93bf-bfaa6be2d1c0-config\") pod \"authentication-operator-69f744f599-jdjcq\" (UID: \"a7e5d16a-d45d-4b40-93bf-bfaa6be2d1c0\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jdjcq" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.596114 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1afc0f4c-e02d-4a70-aaba-e761e8c04eee-client-ca\") pod \"controller-manager-879f6c89f-lndnt\" (UID: \"1afc0f4c-e02d-4a70-aaba-e761e8c04eee\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lndnt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.596132 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/437b5573-a342-4383-ba60-be0e3ccba839-node-bootstrap-token\") pod \"machine-config-server-c9qdp\" (UID: \"437b5573-a342-4383-ba60-be0e3ccba839\") " pod="openshift-machine-config-operator/machine-config-server-c9qdp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.596151 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkngl\" (UniqueName: \"kubernetes.io/projected/5c244eff-aada-44f3-b250-96878a3400c4-kube-api-access-nkngl\") pod \"etcd-operator-b45778765-27vjc\" (UID: \"5c244eff-aada-44f3-b250-96878a3400c4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-27vjc" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.596176 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkl7p\" (UniqueName: \"kubernetes.io/projected/359660cd-b412-4640-bedf-993e976e7b3c-kube-api-access-rkl7p\") pod \"openshift-apiserver-operator-796bbdcf4f-88lkr\" (UID: \"359660cd-b412-4640-bedf-993e976e7b3c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-88lkr" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.596205 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6clr\" (UniqueName: \"kubernetes.io/projected/37a77f41-5dbf-4842-9e77-83dc22b50f4a-kube-api-access-w6clr\") pod \"migrator-59844c95c7-qdmxz\" (UID: \"37a77f41-5dbf-4842-9e77-83dc22b50f4a\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qdmxz" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.596230 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7-config-volume\") pod \"collect-profiles-29490780-8q6q4\" (UID: \"7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-8q6q4" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.596253 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vq4z\" (UniqueName: \"kubernetes.io/projected/a7e5d16a-d45d-4b40-93bf-bfaa6be2d1c0-kube-api-access-7vq4z\") pod \"authentication-operator-69f744f599-jdjcq\" (UID: \"a7e5d16a-d45d-4b40-93bf-bfaa6be2d1c0\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jdjcq" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.596271 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/42c0e428-821f-45a1-85a7-54ebdb81ef1c-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cl895\" (UID: \"42c0e428-821f-45a1-85a7-54ebdb81ef1c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cl895" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.596293 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8qsf\" (UniqueName: \"kubernetes.io/projected/2d37efbf-d18f-486b-9b43-bc4d181af4ca-kube-api-access-b8qsf\") pod \"marketplace-operator-79b997595-wvttb\" (UID: \"2d37efbf-d18f-486b-9b43-bc4d181af4ca\") " pod="openshift-marketplace/marketplace-operator-79b997595-wvttb" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.596317 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hc4dg\" (UniqueName: \"kubernetes.io/projected/abbeffe1-cfd5-4476-9a8e-2ab5b4869444-kube-api-access-hc4dg\") pod \"catalog-operator-68c6474976-nn46h\" (UID: \"abbeffe1-cfd5-4476-9a8e-2ab5b4869444\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nn46h" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.597031 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7e5d16a-d45d-4b40-93bf-bfaa6be2d1c0-config\") pod \"authentication-operator-69f744f599-jdjcq\" (UID: \"a7e5d16a-d45d-4b40-93bf-bfaa6be2d1c0\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jdjcq" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.597154 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf1f11c8-17b8-49b7-b12d-92891f478222-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-7p5jt\" (UID: \"bf1f11c8-17b8-49b7-b12d-92891f478222\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7p5jt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.597412 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/831dc87e-8e14-43d3-a36e-dc7679041ae5-config\") pod \"console-operator-58897d9998-4pbj2\" (UID: \"831dc87e-8e14-43d3-a36e-dc7679041ae5\") " pod="openshift-console-operator/console-operator-58897d9998-4pbj2" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.597477 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b28404ed-2e71-4b3f-9140-35ee89dbc8f2-oauth-serving-cert\") pod \"console-f9d7485db-6qgnn\" (UID: \"b28404ed-2e71-4b3f-9140-35ee89dbc8f2\") " pod="openshift-console/console-f9d7485db-6qgnn" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.596515 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6d331bd-2db3-4319-9f5c-db56d408d9e3-config\") pod \"apiserver-76f77b778f-6rlxp\" (UID: \"a6d331bd-2db3-4319-9f5c-db56d408d9e3\") " pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.597660 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b28404ed-2e71-4b3f-9140-35ee89dbc8f2-trusted-ca-bundle\") pod \"console-f9d7485db-6qgnn\" (UID: \"b28404ed-2e71-4b3f-9140-35ee89dbc8f2\") " pod="openshift-console/console-f9d7485db-6qgnn" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.597693 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgzl5\" (UniqueName: \"kubernetes.io/projected/c8657575-cd22-4ebc-ae9d-4174366985d3-kube-api-access-fgzl5\") pod \"csi-hostpathplugin-vfm8t\" (UID: \"c8657575-cd22-4ebc-ae9d-4174366985d3\") " pod="hostpath-provisioner/csi-hostpathplugin-vfm8t" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.597758 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ddc2e6b7-5582-4579-bf2c-ed165b74c91a-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-vmwvg\" (UID: \"ddc2e6b7-5582-4579-bf2c-ed165b74c91a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vmwvg" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.597788 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1afc0f4c-e02d-4a70-aaba-e761e8c04eee-client-ca\") pod \"controller-manager-879f6c89f-lndnt\" (UID: \"1afc0f4c-e02d-4a70-aaba-e761e8c04eee\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lndnt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.597883 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/c8657575-cd22-4ebc-ae9d-4174366985d3-csi-data-dir\") pod \"csi-hostpathplugin-vfm8t\" (UID: \"c8657575-cd22-4ebc-ae9d-4174366985d3\") " pod="hostpath-provisioner/csi-hostpathplugin-vfm8t" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.597910 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a4d83db5-776f-4e95-a6fa-b194344f9819-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-2sfhr\" (UID: \"a4d83db5-776f-4e95-a6fa-b194344f9819\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2sfhr" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.597939 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/54a246a2-f674-4735-b295-b56699ece95b-auth-proxy-config\") pod \"machine-approver-56656f9798-962cr\" (UID: \"54a246a2-f674-4735-b295-b56699ece95b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-962cr" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.597957 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6fxv\" (UniqueName: \"kubernetes.io/projected/f9b8f57e-00b9-4355-ace2-0319d320d208-kube-api-access-d6fxv\") pod \"packageserver-d55dfcdfc-mr7cp\" (UID: \"f9b8f57e-00b9-4355-ace2-0319d320d208\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mr7cp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.597974 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/f9b8f57e-00b9-4355-ace2-0319d320d208-tmpfs\") pod \"packageserver-d55dfcdfc-mr7cp\" (UID: \"f9b8f57e-00b9-4355-ace2-0319d320d208\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mr7cp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.597989 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/5c244eff-aada-44f3-b250-96878a3400c4-etcd-service-ca\") pod \"etcd-operator-b45778765-27vjc\" (UID: \"5c244eff-aada-44f3-b250-96878a3400c4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-27vjc" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.598006 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a7e5d16a-d45d-4b40-93bf-bfaa6be2d1c0-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-jdjcq\" (UID: \"a7e5d16a-d45d-4b40-93bf-bfaa6be2d1c0\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jdjcq" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.598024 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77a97acb-2908-48fb-8bcd-0647f3e90160-config\") pod \"machine-api-operator-5694c8668f-7xb2b\" (UID: \"77a97acb-2908-48fb-8bcd-0647f3e90160\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7xb2b" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.598039 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2274g\" (UniqueName: \"kubernetes.io/projected/54a246a2-f674-4735-b295-b56699ece95b-kube-api-access-2274g\") pod \"machine-approver-56656f9798-962cr\" (UID: \"54a246a2-f674-4735-b295-b56699ece95b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-962cr" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.598057 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4z5z\" (UniqueName: \"kubernetes.io/projected/05d74105-0ecd-41ac-9001-8b21b0fd6ba4-kube-api-access-m4z5z\") pod \"openshift-controller-manager-operator-756b6f6bc6-l9nqd\" (UID: \"05d74105-0ecd-41ac-9001-8b21b0fd6ba4\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l9nqd" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.598076 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b28404ed-2e71-4b3f-9140-35ee89dbc8f2-console-oauth-config\") pod \"console-f9d7485db-6qgnn\" (UID: \"b28404ed-2e71-4b3f-9140-35ee89dbc8f2\") " pod="openshift-console/console-f9d7485db-6qgnn" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.598094 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-cb8nk\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.598110 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b28404ed-2e71-4b3f-9140-35ee89dbc8f2-service-ca\") pod \"console-f9d7485db-6qgnn\" (UID: \"b28404ed-2e71-4b3f-9140-35ee89dbc8f2\") " pod="openshift-console/console-f9d7485db-6qgnn" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.598131 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/beb6f283-75cb-4184-b985-4e6c095feca1-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-ddghz\" (UID: \"beb6f283-75cb-4184-b985-4e6c095feca1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ddghz" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.598156 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2d37efbf-d18f-486b-9b43-bc4d181af4ca-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wvttb\" (UID: \"2d37efbf-d18f-486b-9b43-bc4d181af4ca\") " pod="openshift-marketplace/marketplace-operator-79b997595-wvttb" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.598178 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/550752e4-a1d9-46f4-9118-9e9919b2fe6b-serving-cert\") pod \"service-ca-operator-777779d784-cjzsq\" (UID: \"550752e4-a1d9-46f4-9118-9e9919b2fe6b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cjzsq" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.598196 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ac10f013-cd1f-47e0-8f1c-5ff4e6e75784-config-volume\") pod \"dns-default-dgcqn\" (UID: \"ac10f013-cd1f-47e0-8f1c-5ff4e6e75784\") " pod="openshift-dns/dns-default-dgcqn" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.598226 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/69008ed1-f3e5-400d-852f-adbcd94199f6-audit-policies\") pod \"oauth-openshift-558db77b4-cb8nk\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.598241 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-cb8nk\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.598255 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddc2e6b7-5582-4579-bf2c-ed165b74c91a-config\") pod \"kube-apiserver-operator-766d6c64bb-vmwvg\" (UID: \"ddc2e6b7-5582-4579-bf2c-ed165b74c91a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vmwvg" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.598269 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/77a97acb-2908-48fb-8bcd-0647f3e90160-images\") pod \"machine-api-operator-5694c8668f-7xb2b\" (UID: \"77a97acb-2908-48fb-8bcd-0647f3e90160\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7xb2b" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.598300 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/77a97acb-2908-48fb-8bcd-0647f3e90160-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-7xb2b\" (UID: \"77a97acb-2908-48fb-8bcd-0647f3e90160\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7xb2b" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.598317 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/81c2f96b-55e0-483b-b72c-df7e156e9218-serving-cert\") pod \"apiserver-7bbb656c7d-6cghs\" (UID: \"81c2f96b-55e0-483b-b72c-df7e156e9218\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6cghs" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.598332 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-cb8nk\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.598350 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/c8657575-cd22-4ebc-ae9d-4174366985d3-mountpoint-dir\") pod \"csi-hostpathplugin-vfm8t\" (UID: \"c8657575-cd22-4ebc-ae9d-4174366985d3\") " pod="hostpath-provisioner/csi-hostpathplugin-vfm8t" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.598366 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/c8657575-cd22-4ebc-ae9d-4174366985d3-plugins-dir\") pod \"csi-hostpathplugin-vfm8t\" (UID: \"c8657575-cd22-4ebc-ae9d-4174366985d3\") " pod="hostpath-provisioner/csi-hostpathplugin-vfm8t" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.598385 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c244eff-aada-44f3-b250-96878a3400c4-config\") pod \"etcd-operator-b45778765-27vjc\" (UID: \"5c244eff-aada-44f3-b250-96878a3400c4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-27vjc" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.598476 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6d331bd-2db3-4319-9f5c-db56d408d9e3-config\") pod \"apiserver-76f77b778f-6rlxp\" (UID: \"a6d331bd-2db3-4319-9f5c-db56d408d9e3\") " pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.599241 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/54a246a2-f674-4735-b295-b56699ece95b-auth-proxy-config\") pod \"machine-approver-56656f9798-962cr\" (UID: \"54a246a2-f674-4735-b295-b56699ece95b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-962cr" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.599268 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fa11789e-7a2a-4dbf-85ca-c20a9d64a1f4-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-6snv6\" (UID: \"fa11789e-7a2a-4dbf-85ca-c20a9d64a1f4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6snv6" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.599332 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kf2b2\" (UniqueName: \"kubernetes.io/projected/69008ed1-f3e5-400d-852f-adbcd94199f6-kube-api-access-kf2b2\") pod \"oauth-openshift-558db77b4-cb8nk\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.599357 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a6d331bd-2db3-4319-9f5c-db56d408d9e3-node-pullsecrets\") pod \"apiserver-76f77b778f-6rlxp\" (UID: \"a6d331bd-2db3-4319-9f5c-db56d408d9e3\") " pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.599377 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1afc0f4c-e02d-4a70-aaba-e761e8c04eee-serving-cert\") pod \"controller-manager-879f6c89f-lndnt\" (UID: \"1afc0f4c-e02d-4a70-aaba-e761e8c04eee\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lndnt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.599396 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/a6d331bd-2db3-4319-9f5c-db56d408d9e3-etcd-serving-ca\") pod \"apiserver-76f77b778f-6rlxp\" (UID: \"a6d331bd-2db3-4319-9f5c-db56d408d9e3\") " pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.599416 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2hpl\" (UniqueName: \"kubernetes.io/projected/a4d83db5-776f-4e95-a6fa-b194344f9819-kube-api-access-t2hpl\") pod \"machine-config-controller-84d6567774-2sfhr\" (UID: \"a4d83db5-776f-4e95-a6fa-b194344f9819\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2sfhr" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.599436 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-cb8nk\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.599459 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/129a0b30-7132-4e3c-ab84-208cae7cb2f2-proxy-tls\") pod \"machine-config-operator-74547568cd-zzxln\" (UID: \"129a0b30-7132-4e3c-ab84-208cae7cb2f2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-zzxln" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.599521 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1afc0f4c-e02d-4a70-aaba-e761e8c04eee-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-lndnt\" (UID: \"1afc0f4c-e02d-4a70-aaba-e761e8c04eee\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lndnt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.599573 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a7e5d16a-d45d-4b40-93bf-bfaa6be2d1c0-serving-cert\") pod \"authentication-operator-69f744f599-jdjcq\" (UID: \"a7e5d16a-d45d-4b40-93bf-bfaa6be2d1c0\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jdjcq" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.599578 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvkhn\" (UniqueName: \"kubernetes.io/projected/b28404ed-2e71-4b3f-9140-35ee89dbc8f2-kube-api-access-dvkhn\") pod \"console-f9d7485db-6qgnn\" (UID: \"b28404ed-2e71-4b3f-9140-35ee89dbc8f2\") " pod="openshift-console/console-f9d7485db-6qgnn" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.599635 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/77a97acb-2908-48fb-8bcd-0647f3e90160-images\") pod \"machine-api-operator-5694c8668f-7xb2b\" (UID: \"77a97acb-2908-48fb-8bcd-0647f3e90160\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7xb2b" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.599661 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a6d331bd-2db3-4319-9f5c-db56d408d9e3-node-pullsecrets\") pod \"apiserver-76f77b778f-6rlxp\" (UID: \"a6d331bd-2db3-4319-9f5c-db56d408d9e3\") " pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.600109 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-cb8nk\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.600369 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/5c244eff-aada-44f3-b250-96878a3400c4-etcd-service-ca\") pod \"etcd-operator-b45778765-27vjc\" (UID: \"5c244eff-aada-44f3-b250-96878a3400c4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-27vjc" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.600550 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c244eff-aada-44f3-b250-96878a3400c4-config\") pod \"etcd-operator-b45778765-27vjc\" (UID: \"5c244eff-aada-44f3-b250-96878a3400c4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-27vjc" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.600634 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b28404ed-2e71-4b3f-9140-35ee89dbc8f2-service-ca\") pod \"console-f9d7485db-6qgnn\" (UID: \"b28404ed-2e71-4b3f-9140-35ee89dbc8f2\") " pod="openshift-console/console-f9d7485db-6qgnn" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.600806 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77a97acb-2908-48fb-8bcd-0647f3e90160-config\") pod \"machine-api-operator-5694c8668f-7xb2b\" (UID: \"77a97acb-2908-48fb-8bcd-0647f3e90160\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7xb2b" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.600924 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a7e5d16a-d45d-4b40-93bf-bfaa6be2d1c0-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-jdjcq\" (UID: \"a7e5d16a-d45d-4b40-93bf-bfaa6be2d1c0\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jdjcq" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.601134 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a6d331bd-2db3-4319-9f5c-db56d408d9e3-serving-cert\") pod \"apiserver-76f77b778f-6rlxp\" (UID: \"a6d331bd-2db3-4319-9f5c-db56d408d9e3\") " pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.601268 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/a6d331bd-2db3-4319-9f5c-db56d408d9e3-etcd-serving-ca\") pod \"apiserver-76f77b778f-6rlxp\" (UID: \"a6d331bd-2db3-4319-9f5c-db56d408d9e3\") " pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.601448 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1afc0f4c-e02d-4a70-aaba-e761e8c04eee-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-lndnt\" (UID: \"1afc0f4c-e02d-4a70-aaba-e761e8c04eee\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lndnt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.601731 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-cb8nk\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.601871 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/f9b8f57e-00b9-4355-ace2-0319d320d208-tmpfs\") pod \"packageserver-d55dfcdfc-mr7cp\" (UID: \"f9b8f57e-00b9-4355-ace2-0319d320d208\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mr7cp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.602044 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-cb8nk\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.602044 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/033cb12f-278f-431a-8104-519db9a3152f-signing-cabundle\") pod \"service-ca-9c57cc56f-gz7kg\" (UID: \"033cb12f-278f-431a-8104-519db9a3152f\") " pod="openshift-service-ca/service-ca-9c57cc56f-gz7kg" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.602146 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5scg\" (UniqueName: \"kubernetes.io/projected/033cb12f-278f-431a-8104-519db9a3152f-kube-api-access-s5scg\") pod \"service-ca-9c57cc56f-gz7kg\" (UID: \"033cb12f-278f-431a-8104-519db9a3152f\") " pod="openshift-service-ca/service-ca-9c57cc56f-gz7kg" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.602176 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a6d331bd-2db3-4319-9f5c-db56d408d9e3-trusted-ca-bundle\") pod \"apiserver-76f77b778f-6rlxp\" (UID: \"a6d331bd-2db3-4319-9f5c-db56d408d9e3\") " pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.602199 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/359660cd-b412-4640-bedf-993e976e7b3c-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-88lkr\" (UID: \"359660cd-b412-4640-bedf-993e976e7b3c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-88lkr" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.602223 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05d74105-0ecd-41ac-9001-8b21b0fd6ba4-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-l9nqd\" (UID: \"05d74105-0ecd-41ac-9001-8b21b0fd6ba4\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l9nqd" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.602271 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5c244eff-aada-44f3-b250-96878a3400c4-etcd-client\") pod \"etcd-operator-b45778765-27vjc\" (UID: \"5c244eff-aada-44f3-b250-96878a3400c4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-27vjc" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.602301 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ac10f013-cd1f-47e0-8f1c-5ff4e6e75784-metrics-tls\") pod \"dns-default-dgcqn\" (UID: \"ac10f013-cd1f-47e0-8f1c-5ff4e6e75784\") " pod="openshift-dns/dns-default-dgcqn" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.602329 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-cb8nk\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.602356 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b28404ed-2e71-4b3f-9140-35ee89dbc8f2-console-config\") pod \"console-f9d7485db-6qgnn\" (UID: \"b28404ed-2e71-4b3f-9140-35ee89dbc8f2\") " pod="openshift-console/console-f9d7485db-6qgnn" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.602388 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/004316da-16cd-49ab-b14d-282c28da6fad-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-8m4l6\" (UID: \"004316da-16cd-49ab-b14d-282c28da6fad\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8m4l6" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.602418 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nr9jp\" (UniqueName: \"kubernetes.io/projected/004316da-16cd-49ab-b14d-282c28da6fad-kube-api-access-nr9jp\") pod \"package-server-manager-789f6589d5-8m4l6\" (UID: \"004316da-16cd-49ab-b14d-282c28da6fad\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8m4l6" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.602441 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/77a97acb-2908-48fb-8bcd-0647f3e90160-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-7xb2b\" (UID: \"77a97acb-2908-48fb-8bcd-0647f3e90160\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7xb2b" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.602554 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-cb8nk\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.602447 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-cb8nk\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.602622 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/550752e4-a1d9-46f4-9118-9e9919b2fe6b-config\") pod \"service-ca-operator-777779d784-cjzsq\" (UID: \"550752e4-a1d9-46f4-9118-9e9919b2fe6b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cjzsq" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.602656 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-cb8nk\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.602677 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/73de6ef2-e139-4185-9f56-9db885734ffe-bound-sa-token\") pod \"ingress-operator-5b745b69d9-58fcz\" (UID: \"73de6ef2-e139-4185-9f56-9db885734ffe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-58fcz" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.602697 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42c0e428-821f-45a1-85a7-54ebdb81ef1c-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cl895\" (UID: \"42c0e428-821f-45a1-85a7-54ebdb81ef1c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cl895" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.602716 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/81c2f96b-55e0-483b-b72c-df7e156e9218-audit-dir\") pod \"apiserver-7bbb656c7d-6cghs\" (UID: \"81c2f96b-55e0-483b-b72c-df7e156e9218\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6cghs" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.602733 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqhmb\" (UniqueName: \"kubernetes.io/projected/81c2f96b-55e0-483b-b72c-df7e156e9218-kube-api-access-rqhmb\") pod \"apiserver-7bbb656c7d-6cghs\" (UID: \"81c2f96b-55e0-483b-b72c-df7e156e9218\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6cghs" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.602759 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/149e3000-35d7-47bd-83f0-00ab5e0736c2-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-q7j7b\" (UID: \"149e3000-35d7-47bd-83f0-00ab5e0736c2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-q7j7b" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.602776 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/73de6ef2-e139-4185-9f56-9db885734ffe-metrics-tls\") pod \"ingress-operator-5b745b69d9-58fcz\" (UID: \"73de6ef2-e139-4185-9f56-9db885734ffe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-58fcz" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.602793 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a4d83db5-776f-4e95-a6fa-b194344f9819-proxy-tls\") pod \"machine-config-controller-84d6567774-2sfhr\" (UID: \"a4d83db5-776f-4e95-a6fa-b194344f9819\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2sfhr" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.602812 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/a6d331bd-2db3-4319-9f5c-db56d408d9e3-image-import-ca\") pod \"apiserver-76f77b778f-6rlxp\" (UID: \"a6d331bd-2db3-4319-9f5c-db56d408d9e3\") " pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.602834 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6v8bp\" (UniqueName: \"kubernetes.io/projected/ac10f013-cd1f-47e0-8f1c-5ff4e6e75784-kube-api-access-6v8bp\") pod \"dns-default-dgcqn\" (UID: \"ac10f013-cd1f-47e0-8f1c-5ff4e6e75784\") " pod="openshift-dns/dns-default-dgcqn" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.602859 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a1546392-4a69-4b12-8d7e-97450b73b7ca-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-rrhjv\" (UID: \"a1546392-4a69-4b12-8d7e-97450b73b7ca\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rrhjv" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.602880 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qcw5\" (UniqueName: \"kubernetes.io/projected/bf1f11c8-17b8-49b7-b12d-92891f478222-kube-api-access-9qcw5\") pod \"cluster-image-registry-operator-dc59b4c8b-7p5jt\" (UID: \"bf1f11c8-17b8-49b7-b12d-92891f478222\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7p5jt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.602899 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2wtm\" (UniqueName: \"kubernetes.io/projected/0c1af7db-aa80-4cb0-a9cb-5afdf677f28c-kube-api-access-v2wtm\") pod \"cluster-samples-operator-665b6dd947-4w5bf\" (UID: \"0c1af7db-aa80-4cb0-a9cb-5afdf677f28c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4w5bf" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.602959 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/69008ed1-f3e5-400d-852f-adbcd94199f6-audit-dir\") pod \"oauth-openshift-558db77b4-cb8nk\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.602980 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zchmc\" (UniqueName: \"kubernetes.io/projected/550752e4-a1d9-46f4-9118-9e9919b2fe6b-kube-api-access-zchmc\") pod \"service-ca-operator-777779d784-cjzsq\" (UID: \"550752e4-a1d9-46f4-9118-9e9919b2fe6b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cjzsq" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.603011 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f9b8f57e-00b9-4355-ace2-0319d320d208-webhook-cert\") pod \"packageserver-d55dfcdfc-mr7cp\" (UID: \"f9b8f57e-00b9-4355-ace2-0319d320d208\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mr7cp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.603040 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/abbeffe1-cfd5-4476-9a8e-2ab5b4869444-srv-cert\") pod \"catalog-operator-68c6474976-nn46h\" (UID: \"abbeffe1-cfd5-4476-9a8e-2ab5b4869444\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nn46h" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.603047 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/69008ed1-f3e5-400d-852f-adbcd94199f6-audit-policies\") pod \"oauth-openshift-558db77b4-cb8nk\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.603065 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/0c1af7db-aa80-4cb0-a9cb-5afdf677f28c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-4w5bf\" (UID: \"0c1af7db-aa80-4cb0-a9cb-5afdf677f28c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4w5bf" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.603097 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/73de6ef2-e139-4185-9f56-9db885734ffe-trusted-ca\") pod \"ingress-operator-5b745b69d9-58fcz\" (UID: \"73de6ef2-e139-4185-9f56-9db885734ffe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-58fcz" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.603124 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgb6k\" (UniqueName: \"kubernetes.io/projected/94291fa4-24a5-499e-8143-89c8784d9284-kube-api-access-hgb6k\") pod \"downloads-7954f5f757-7l927\" (UID: \"94291fa4-24a5-499e-8143-89c8784d9284\") " pod="openshift-console/downloads-7954f5f757-7l927" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.603157 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ddc2e6b7-5582-4579-bf2c-ed165b74c91a-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-vmwvg\" (UID: \"ddc2e6b7-5582-4579-bf2c-ed165b74c91a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vmwvg" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.603177 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/81c2f96b-55e0-483b-b72c-df7e156e9218-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-6cghs\" (UID: \"81c2f96b-55e0-483b-b72c-df7e156e9218\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6cghs" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.603197 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjhr5\" (UniqueName: \"kubernetes.io/projected/beb6f283-75cb-4184-b985-4e6c095feca1-kube-api-access-mjhr5\") pod \"multus-admission-controller-857f4d67dd-ddghz\" (UID: \"beb6f283-75cb-4184-b985-4e6c095feca1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ddghz" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.603221 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05d74105-0ecd-41ac-9001-8b21b0fd6ba4-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-l9nqd\" (UID: \"05d74105-0ecd-41ac-9001-8b21b0fd6ba4\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l9nqd" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.603245 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7lk4\" (UniqueName: \"kubernetes.io/projected/85f05bd5-ff83-4d29-9531-ab3499088095-kube-api-access-x7lk4\") pod \"router-default-5444994796-h9b2g\" (UID: \"85f05bd5-ff83-4d29-9531-ab3499088095\") " pod="openshift-ingress/router-default-5444994796-h9b2g" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.603277 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/149e3000-35d7-47bd-83f0-00ab5e0736c2-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-q7j7b\" (UID: \"149e3000-35d7-47bd-83f0-00ab5e0736c2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-q7j7b" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.603325 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/85f05bd5-ff83-4d29-9531-ab3499088095-metrics-certs\") pod \"router-default-5444994796-h9b2g\" (UID: \"85f05bd5-ff83-4d29-9531-ab3499088095\") " pod="openshift-ingress/router-default-5444994796-h9b2g" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.603356 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/54a246a2-f674-4735-b295-b56699ece95b-machine-approver-tls\") pod \"machine-approver-56656f9798-962cr\" (UID: \"54a246a2-f674-4735-b295-b56699ece95b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-962cr" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.603379 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/17a72e73-4d54-4a29-a85a-ecb1aff30d10-srv-cert\") pod \"olm-operator-6b444d44fb-k662z\" (UID: \"17a72e73-4d54-4a29-a85a-ecb1aff30d10\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k662z" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.603399 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vxn2\" (UniqueName: \"kubernetes.io/projected/437b5573-a342-4383-ba60-be0e3ccba839-kube-api-access-9vxn2\") pod \"machine-config-server-c9qdp\" (UID: \"437b5573-a342-4383-ba60-be0e3ccba839\") " pod="openshift-machine-config-operator/machine-config-server-c9qdp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.603431 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a6d331bd-2db3-4319-9f5c-db56d408d9e3-etcd-client\") pod \"apiserver-76f77b778f-6rlxp\" (UID: \"a6d331bd-2db3-4319-9f5c-db56d408d9e3\") " pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.603454 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-cb8nk\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.603475 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/437b5573-a342-4383-ba60-be0e3ccba839-certs\") pod \"machine-config-server-c9qdp\" (UID: \"437b5573-a342-4383-ba60-be0e3ccba839\") " pod="openshift-machine-config-operator/machine-config-server-c9qdp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.603499 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/033cb12f-278f-431a-8104-519db9a3152f-signing-key\") pod \"service-ca-9c57cc56f-gz7kg\" (UID: \"033cb12f-278f-431a-8104-519db9a3152f\") " pod="openshift-service-ca/service-ca-9c57cc56f-gz7kg" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.603547 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/831dc87e-8e14-43d3-a36e-dc7679041ae5-trusted-ca\") pod \"console-operator-58897d9998-4pbj2\" (UID: \"831dc87e-8e14-43d3-a36e-dc7679041ae5\") " pod="openshift-console-operator/console-operator-58897d9998-4pbj2" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.603580 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c475g\" (UniqueName: \"kubernetes.io/projected/77a97acb-2908-48fb-8bcd-0647f3e90160-kube-api-access-c475g\") pod \"machine-api-operator-5694c8668f-7xb2b\" (UID: \"77a97acb-2908-48fb-8bcd-0647f3e90160\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7xb2b" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.603604 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/85f05bd5-ff83-4d29-9531-ab3499088095-stats-auth\") pod \"router-default-5444994796-h9b2g\" (UID: \"85f05bd5-ff83-4d29-9531-ab3499088095\") " pod="openshift-ingress/router-default-5444994796-h9b2g" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.603631 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kkkd\" (UniqueName: \"kubernetes.io/projected/129a0b30-7132-4e3c-ab84-208cae7cb2f2-kube-api-access-6kkkd\") pod \"machine-config-operator-74547568cd-zzxln\" (UID: \"129a0b30-7132-4e3c-ab84-208cae7cb2f2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-zzxln" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.603660 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2wnd\" (UniqueName: \"kubernetes.io/projected/831dc87e-8e14-43d3-a36e-dc7679041ae5-kube-api-access-d2wnd\") pod \"console-operator-58897d9998-4pbj2\" (UID: \"831dc87e-8e14-43d3-a36e-dc7679041ae5\") " pod="openshift-console-operator/console-operator-58897d9998-4pbj2" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.603685 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/17a72e73-4d54-4a29-a85a-ecb1aff30d10-profile-collector-cert\") pod \"olm-operator-6b444d44fb-k662z\" (UID: \"17a72e73-4d54-4a29-a85a-ecb1aff30d10\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k662z" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.603709 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2d37efbf-d18f-486b-9b43-bc4d181af4ca-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wvttb\" (UID: \"2d37efbf-d18f-486b-9b43-bc4d181af4ca\") " pod="openshift-marketplace/marketplace-operator-79b997595-wvttb" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.603734 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b28404ed-2e71-4b3f-9140-35ee89dbc8f2-console-serving-cert\") pod \"console-f9d7485db-6qgnn\" (UID: \"b28404ed-2e71-4b3f-9140-35ee89dbc8f2\") " pod="openshift-console/console-f9d7485db-6qgnn" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.603758 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f9b8f57e-00b9-4355-ace2-0319d320d208-apiservice-cert\") pod \"packageserver-d55dfcdfc-mr7cp\" (UID: \"f9b8f57e-00b9-4355-ace2-0319d320d208\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mr7cp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.603788 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/359660cd-b412-4640-bedf-993e976e7b3c-config\") pod \"openshift-apiserver-operator-796bbdcf4f-88lkr\" (UID: \"359660cd-b412-4640-bedf-993e976e7b3c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-88lkr" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.603813 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-cb8nk\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.603840 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdxxq\" (UniqueName: \"kubernetes.io/projected/73de6ef2-e139-4185-9f56-9db885734ffe-kube-api-access-hdxxq\") pod \"ingress-operator-5b745b69d9-58fcz\" (UID: \"73de6ef2-e139-4185-9f56-9db885734ffe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-58fcz" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.603864 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/129a0b30-7132-4e3c-ab84-208cae7cb2f2-auth-proxy-config\") pod \"machine-config-operator-74547568cd-zzxln\" (UID: \"129a0b30-7132-4e3c-ab84-208cae7cb2f2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-zzxln" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.603891 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/5c244eff-aada-44f3-b250-96878a3400c4-etcd-ca\") pod \"etcd-operator-b45778765-27vjc\" (UID: \"5c244eff-aada-44f3-b250-96878a3400c4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-27vjc" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.603912 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c244eff-aada-44f3-b250-96878a3400c4-serving-cert\") pod \"etcd-operator-b45778765-27vjc\" (UID: \"5c244eff-aada-44f3-b250-96878a3400c4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-27vjc" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.603933 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7-secret-volume\") pod \"collect-profiles-29490780-8q6q4\" (UID: \"7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-8q6q4" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.603955 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/129a0b30-7132-4e3c-ab84-208cae7cb2f2-images\") pod \"machine-config-operator-74547568cd-zzxln\" (UID: \"129a0b30-7132-4e3c-ab84-208cae7cb2f2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-zzxln" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.603985 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/c8657575-cd22-4ebc-ae9d-4174366985d3-registration-dir\") pod \"csi-hostpathplugin-vfm8t\" (UID: \"c8657575-cd22-4ebc-ae9d-4174366985d3\") " pod="hostpath-provisioner/csi-hostpathplugin-vfm8t" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.604014 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/81c2f96b-55e0-483b-b72c-df7e156e9218-encryption-config\") pod \"apiserver-7bbb656c7d-6cghs\" (UID: \"81c2f96b-55e0-483b-b72c-df7e156e9218\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6cghs" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.604038 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxgfj\" (UniqueName: \"kubernetes.io/projected/7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7-kube-api-access-kxgfj\") pod \"collect-profiles-29490780-8q6q4\" (UID: \"7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-8q6q4" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.604062 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-cb8nk\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.604086 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/c8657575-cd22-4ebc-ae9d-4174366985d3-socket-dir\") pod \"csi-hostpathplugin-vfm8t\" (UID: \"c8657575-cd22-4ebc-ae9d-4174366985d3\") " pod="hostpath-provisioner/csi-hostpathplugin-vfm8t" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.604103 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fe6baed-ab97-4d8a-8be2-6f00f9698136-config\") pod \"route-controller-manager-6576b87f9c-fpqvc\" (UID: \"5fe6baed-ab97-4d8a-8be2-6f00f9698136\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpqvc" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.604112 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/a6d331bd-2db3-4319-9f5c-db56d408d9e3-image-import-ca\") pod \"apiserver-76f77b778f-6rlxp\" (UID: \"a6d331bd-2db3-4319-9f5c-db56d408d9e3\") " pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.604108 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/a6d331bd-2db3-4319-9f5c-db56d408d9e3-encryption-config\") pod \"apiserver-76f77b778f-6rlxp\" (UID: \"a6d331bd-2db3-4319-9f5c-db56d408d9e3\") " pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.604177 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf1f11c8-17b8-49b7-b12d-92891f478222-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-7p5jt\" (UID: \"bf1f11c8-17b8-49b7-b12d-92891f478222\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7p5jt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.604352 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/69008ed1-f3e5-400d-852f-adbcd94199f6-audit-dir\") pod \"oauth-openshift-558db77b4-cb8nk\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.604698 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-cb8nk\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.604707 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/81c2f96b-55e0-483b-b72c-df7e156e9218-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-6cghs\" (UID: \"81c2f96b-55e0-483b-b72c-df7e156e9218\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6cghs" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.605064 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b28404ed-2e71-4b3f-9140-35ee89dbc8f2-console-config\") pod \"console-f9d7485db-6qgnn\" (UID: \"b28404ed-2e71-4b3f-9140-35ee89dbc8f2\") " pod="openshift-console/console-f9d7485db-6qgnn" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.605478 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1afc0f4c-e02d-4a70-aaba-e761e8c04eee-serving-cert\") pod \"controller-manager-879f6c89f-lndnt\" (UID: \"1afc0f4c-e02d-4a70-aaba-e761e8c04eee\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lndnt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.605519 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/5c244eff-aada-44f3-b250-96878a3400c4-etcd-ca\") pod \"etcd-operator-b45778765-27vjc\" (UID: \"5c244eff-aada-44f3-b250-96878a3400c4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-27vjc" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.605632 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/81c2f96b-55e0-483b-b72c-df7e156e9218-audit-dir\") pod \"apiserver-7bbb656c7d-6cghs\" (UID: \"81c2f96b-55e0-483b-b72c-df7e156e9218\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6cghs" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.605677 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a6d331bd-2db3-4319-9f5c-db56d408d9e3-trusted-ca-bundle\") pod \"apiserver-76f77b778f-6rlxp\" (UID: \"a6d331bd-2db3-4319-9f5c-db56d408d9e3\") " pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.606400 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/359660cd-b412-4640-bedf-993e976e7b3c-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-88lkr\" (UID: \"359660cd-b412-4640-bedf-993e976e7b3c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-88lkr" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.606934 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5c244eff-aada-44f3-b250-96878a3400c4-etcd-client\") pod \"etcd-operator-b45778765-27vjc\" (UID: \"5c244eff-aada-44f3-b250-96878a3400c4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-27vjc" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.607502 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-cb8nk\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.607783 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-cb8nk\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.608189 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/831dc87e-8e14-43d3-a36e-dc7679041ae5-serving-cert\") pod \"console-operator-58897d9998-4pbj2\" (UID: \"831dc87e-8e14-43d3-a36e-dc7679041ae5\") " pod="openshift-console-operator/console-operator-58897d9998-4pbj2" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.608226 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwmjs\" (UniqueName: \"kubernetes.io/projected/a1546392-4a69-4b12-8d7e-97450b73b7ca-kube-api-access-pwmjs\") pod \"control-plane-machine-set-operator-78cbb6b69f-rrhjv\" (UID: \"a1546392-4a69-4b12-8d7e-97450b73b7ca\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rrhjv" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.608325 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/359660cd-b412-4640-bedf-993e976e7b3c-config\") pod \"openshift-apiserver-operator-796bbdcf4f-88lkr\" (UID: \"359660cd-b412-4640-bedf-993e976e7b3c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-88lkr" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.608392 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/a6d331bd-2db3-4319-9f5c-db56d408d9e3-audit\") pod \"apiserver-76f77b778f-6rlxp\" (UID: \"a6d331bd-2db3-4319-9f5c-db56d408d9e3\") " pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.608430 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/81c2f96b-55e0-483b-b72c-df7e156e9218-etcd-client\") pod \"apiserver-7bbb656c7d-6cghs\" (UID: \"81c2f96b-55e0-483b-b72c-df7e156e9218\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6cghs" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.608436 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/149e3000-35d7-47bd-83f0-00ab5e0736c2-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-q7j7b\" (UID: \"149e3000-35d7-47bd-83f0-00ab5e0736c2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-q7j7b" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.608451 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5fe6baed-ab97-4d8a-8be2-6f00f9698136-serving-cert\") pod \"route-controller-manager-6576b87f9c-fpqvc\" (UID: \"5fe6baed-ab97-4d8a-8be2-6f00f9698136\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpqvc" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.608508 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b28404ed-2e71-4b3f-9140-35ee89dbc8f2-console-serving-cert\") pod \"console-f9d7485db-6qgnn\" (UID: \"b28404ed-2e71-4b3f-9140-35ee89dbc8f2\") " pod="openshift-console/console-f9d7485db-6qgnn" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.608581 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/85f05bd5-ff83-4d29-9531-ab3499088095-default-certificate\") pod \"router-default-5444994796-h9b2g\" (UID: \"85f05bd5-ff83-4d29-9531-ab3499088095\") " pod="openshift-ingress/router-default-5444994796-h9b2g" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.608959 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/831dc87e-8e14-43d3-a36e-dc7679041ae5-trusted-ca\") pod \"console-operator-58897d9998-4pbj2\" (UID: \"831dc87e-8e14-43d3-a36e-dc7679041ae5\") " pod="openshift-console-operator/console-operator-58897d9998-4pbj2" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.609213 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/a6d331bd-2db3-4319-9f5c-db56d408d9e3-audit\") pod \"apiserver-76f77b778f-6rlxp\" (UID: \"a6d331bd-2db3-4319-9f5c-db56d408d9e3\") " pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.609261 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5fe6baed-ab97-4d8a-8be2-6f00f9698136-client-ca\") pod \"route-controller-manager-6576b87f9c-fpqvc\" (UID: \"5fe6baed-ab97-4d8a-8be2-6f00f9698136\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpqvc" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.609364 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/54a246a2-f674-4735-b295-b56699ece95b-machine-approver-tls\") pod \"machine-approver-56656f9798-962cr\" (UID: \"54a246a2-f674-4735-b295-b56699ece95b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-962cr" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.609862 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5fe6baed-ab97-4d8a-8be2-6f00f9698136-client-ca\") pod \"route-controller-manager-6576b87f9c-fpqvc\" (UID: \"5fe6baed-ab97-4d8a-8be2-6f00f9698136\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpqvc" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.609871 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/81c2f96b-55e0-483b-b72c-df7e156e9218-serving-cert\") pod \"apiserver-7bbb656c7d-6cghs\" (UID: \"81c2f96b-55e0-483b-b72c-df7e156e9218\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6cghs" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.610308 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-cb8nk\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.610336 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-cb8nk\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.610618 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf1f11c8-17b8-49b7-b12d-92891f478222-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-7p5jt\" (UID: \"bf1f11c8-17b8-49b7-b12d-92891f478222\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7p5jt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.610854 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/149e3000-35d7-47bd-83f0-00ab5e0736c2-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-q7j7b\" (UID: \"149e3000-35d7-47bd-83f0-00ab5e0736c2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-q7j7b" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.610910 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/0c1af7db-aa80-4cb0-a9cb-5afdf677f28c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-4w5bf\" (UID: \"0c1af7db-aa80-4cb0-a9cb-5afdf677f28c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4w5bf" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.611089 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-cb8nk\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.611939 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/81c2f96b-55e0-483b-b72c-df7e156e9218-etcd-client\") pod \"apiserver-7bbb656c7d-6cghs\" (UID: \"81c2f96b-55e0-483b-b72c-df7e156e9218\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6cghs" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.612040 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-cb8nk\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.612134 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a6d331bd-2db3-4319-9f5c-db56d408d9e3-etcd-client\") pod \"apiserver-76f77b778f-6rlxp\" (UID: \"a6d331bd-2db3-4319-9f5c-db56d408d9e3\") " pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.612298 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/831dc87e-8e14-43d3-a36e-dc7679041ae5-serving-cert\") pod \"console-operator-58897d9998-4pbj2\" (UID: \"831dc87e-8e14-43d3-a36e-dc7679041ae5\") " pod="openshift-console-operator/console-operator-58897d9998-4pbj2" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.612488 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c244eff-aada-44f3-b250-96878a3400c4-serving-cert\") pod \"etcd-operator-b45778765-27vjc\" (UID: \"5c244eff-aada-44f3-b250-96878a3400c4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-27vjc" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.612639 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/81c2f96b-55e0-483b-b72c-df7e156e9218-encryption-config\") pod \"apiserver-7bbb656c7d-6cghs\" (UID: \"81c2f96b-55e0-483b-b72c-df7e156e9218\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6cghs" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.612648 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/a6d331bd-2db3-4319-9f5c-db56d408d9e3-encryption-config\") pod \"apiserver-76f77b778f-6rlxp\" (UID: \"a6d331bd-2db3-4319-9f5c-db56d408d9e3\") " pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.614346 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5fe6baed-ab97-4d8a-8be2-6f00f9698136-serving-cert\") pod \"route-controller-manager-6576b87f9c-fpqvc\" (UID: \"5fe6baed-ab97-4d8a-8be2-6f00f9698136\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpqvc" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.616607 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.636587 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.640299 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b28404ed-2e71-4b3f-9140-35ee89dbc8f2-console-oauth-config\") pod \"console-f9d7485db-6qgnn\" (UID: \"b28404ed-2e71-4b3f-9140-35ee89dbc8f2\") " pod="openshift-console/console-f9d7485db-6qgnn" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.657294 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.679594 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.696220 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.710260 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ac10f013-cd1f-47e0-8f1c-5ff4e6e75784-metrics-tls\") pod \"dns-default-dgcqn\" (UID: \"ac10f013-cd1f-47e0-8f1c-5ff4e6e75784\") " pod="openshift-dns/dns-default-dgcqn" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.710493 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nr9jp\" (UniqueName: \"kubernetes.io/projected/004316da-16cd-49ab-b14d-282c28da6fad-kube-api-access-nr9jp\") pod \"package-server-manager-789f6589d5-8m4l6\" (UID: \"004316da-16cd-49ab-b14d-282c28da6fad\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8m4l6" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.710719 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/004316da-16cd-49ab-b14d-282c28da6fad-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-8m4l6\" (UID: \"004316da-16cd-49ab-b14d-282c28da6fad\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8m4l6" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.710814 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/550752e4-a1d9-46f4-9118-9e9919b2fe6b-config\") pod \"service-ca-operator-777779d784-cjzsq\" (UID: \"550752e4-a1d9-46f4-9118-9e9919b2fe6b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cjzsq" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.710930 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/73de6ef2-e139-4185-9f56-9db885734ffe-bound-sa-token\") pod \"ingress-operator-5b745b69d9-58fcz\" (UID: \"73de6ef2-e139-4185-9f56-9db885734ffe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-58fcz" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.711040 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/73de6ef2-e139-4185-9f56-9db885734ffe-metrics-tls\") pod \"ingress-operator-5b745b69d9-58fcz\" (UID: \"73de6ef2-e139-4185-9f56-9db885734ffe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-58fcz" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.711119 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a4d83db5-776f-4e95-a6fa-b194344f9819-proxy-tls\") pod \"machine-config-controller-84d6567774-2sfhr\" (UID: \"a4d83db5-776f-4e95-a6fa-b194344f9819\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2sfhr" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.711194 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6v8bp\" (UniqueName: \"kubernetes.io/projected/ac10f013-cd1f-47e0-8f1c-5ff4e6e75784-kube-api-access-6v8bp\") pod \"dns-default-dgcqn\" (UID: \"ac10f013-cd1f-47e0-8f1c-5ff4e6e75784\") " pod="openshift-dns/dns-default-dgcqn" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.711262 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a1546392-4a69-4b12-8d7e-97450b73b7ca-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-rrhjv\" (UID: \"a1546392-4a69-4b12-8d7e-97450b73b7ca\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rrhjv" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.711355 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/abbeffe1-cfd5-4476-9a8e-2ab5b4869444-srv-cert\") pod \"catalog-operator-68c6474976-nn46h\" (UID: \"abbeffe1-cfd5-4476-9a8e-2ab5b4869444\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nn46h" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.711466 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zchmc\" (UniqueName: \"kubernetes.io/projected/550752e4-a1d9-46f4-9118-9e9919b2fe6b-kube-api-access-zchmc\") pod \"service-ca-operator-777779d784-cjzsq\" (UID: \"550752e4-a1d9-46f4-9118-9e9919b2fe6b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cjzsq" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.711602 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/73de6ef2-e139-4185-9f56-9db885734ffe-trusted-ca\") pod \"ingress-operator-5b745b69d9-58fcz\" (UID: \"73de6ef2-e139-4185-9f56-9db885734ffe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-58fcz" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.711733 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/17a72e73-4d54-4a29-a85a-ecb1aff30d10-srv-cert\") pod \"olm-operator-6b444d44fb-k662z\" (UID: \"17a72e73-4d54-4a29-a85a-ecb1aff30d10\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k662z" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.711831 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vxn2\" (UniqueName: \"kubernetes.io/projected/437b5573-a342-4383-ba60-be0e3ccba839-kube-api-access-9vxn2\") pod \"machine-config-server-c9qdp\" (UID: \"437b5573-a342-4383-ba60-be0e3ccba839\") " pod="openshift-machine-config-operator/machine-config-server-c9qdp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.711951 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/437b5573-a342-4383-ba60-be0e3ccba839-certs\") pod \"machine-config-server-c9qdp\" (UID: \"437b5573-a342-4383-ba60-be0e3ccba839\") " pod="openshift-machine-config-operator/machine-config-server-c9qdp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.712069 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kkkd\" (UniqueName: \"kubernetes.io/projected/129a0b30-7132-4e3c-ab84-208cae7cb2f2-kube-api-access-6kkkd\") pod \"machine-config-operator-74547568cd-zzxln\" (UID: \"129a0b30-7132-4e3c-ab84-208cae7cb2f2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-zzxln" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.712196 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/17a72e73-4d54-4a29-a85a-ecb1aff30d10-profile-collector-cert\") pod \"olm-operator-6b444d44fb-k662z\" (UID: \"17a72e73-4d54-4a29-a85a-ecb1aff30d10\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k662z" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.712306 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2d37efbf-d18f-486b-9b43-bc4d181af4ca-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wvttb\" (UID: \"2d37efbf-d18f-486b-9b43-bc4d181af4ca\") " pod="openshift-marketplace/marketplace-operator-79b997595-wvttb" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.712412 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdxxq\" (UniqueName: \"kubernetes.io/projected/73de6ef2-e139-4185-9f56-9db885734ffe-kube-api-access-hdxxq\") pod \"ingress-operator-5b745b69d9-58fcz\" (UID: \"73de6ef2-e139-4185-9f56-9db885734ffe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-58fcz" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.712517 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/129a0b30-7132-4e3c-ab84-208cae7cb2f2-auth-proxy-config\") pod \"machine-config-operator-74547568cd-zzxln\" (UID: \"129a0b30-7132-4e3c-ab84-208cae7cb2f2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-zzxln" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.712677 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7-secret-volume\") pod \"collect-profiles-29490780-8q6q4\" (UID: \"7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-8q6q4" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.712873 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/129a0b30-7132-4e3c-ab84-208cae7cb2f2-images\") pod \"machine-config-operator-74547568cd-zzxln\" (UID: \"129a0b30-7132-4e3c-ab84-208cae7cb2f2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-zzxln" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.714433 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxgfj\" (UniqueName: \"kubernetes.io/projected/7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7-kube-api-access-kxgfj\") pod \"collect-profiles-29490780-8q6q4\" (UID: \"7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-8q6q4" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.714639 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/c8657575-cd22-4ebc-ae9d-4174366985d3-registration-dir\") pod \"csi-hostpathplugin-vfm8t\" (UID: \"c8657575-cd22-4ebc-ae9d-4174366985d3\") " pod="hostpath-provisioner/csi-hostpathplugin-vfm8t" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.714748 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/c8657575-cd22-4ebc-ae9d-4174366985d3-socket-dir\") pod \"csi-hostpathplugin-vfm8t\" (UID: \"c8657575-cd22-4ebc-ae9d-4174366985d3\") " pod="hostpath-provisioner/csi-hostpathplugin-vfm8t" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.715048 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwmjs\" (UniqueName: \"kubernetes.io/projected/a1546392-4a69-4b12-8d7e-97450b73b7ca-kube-api-access-pwmjs\") pod \"control-plane-machine-set-operator-78cbb6b69f-rrhjv\" (UID: \"a1546392-4a69-4b12-8d7e-97450b73b7ca\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rrhjv" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.713154 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/129a0b30-7132-4e3c-ab84-208cae7cb2f2-auth-proxy-config\") pod \"machine-config-operator-74547568cd-zzxln\" (UID: \"129a0b30-7132-4e3c-ab84-208cae7cb2f2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-zzxln" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.714946 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/c8657575-cd22-4ebc-ae9d-4174366985d3-registration-dir\") pod \"csi-hostpathplugin-vfm8t\" (UID: \"c8657575-cd22-4ebc-ae9d-4174366985d3\") " pod="hostpath-provisioner/csi-hostpathplugin-vfm8t" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.714987 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/c8657575-cd22-4ebc-ae9d-4174366985d3-socket-dir\") pod \"csi-hostpathplugin-vfm8t\" (UID: \"c8657575-cd22-4ebc-ae9d-4174366985d3\") " pod="hostpath-provisioner/csi-hostpathplugin-vfm8t" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.714065 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/73de6ef2-e139-4185-9f56-9db885734ffe-metrics-tls\") pod \"ingress-operator-5b745b69d9-58fcz\" (UID: \"73de6ef2-e139-4185-9f56-9db885734ffe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-58fcz" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.715700 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/abbeffe1-cfd5-4476-9a8e-2ab5b4869444-profile-collector-cert\") pod \"catalog-operator-68c6474976-nn46h\" (UID: \"abbeffe1-cfd5-4476-9a8e-2ab5b4869444\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nn46h" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.715825 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wxzv\" (UniqueName: \"kubernetes.io/projected/17a72e73-4d54-4a29-a85a-ecb1aff30d10-kube-api-access-9wxzv\") pod \"olm-operator-6b444d44fb-k662z\" (UID: \"17a72e73-4d54-4a29-a85a-ecb1aff30d10\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k662z" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.715936 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/437b5573-a342-4383-ba60-be0e3ccba839-node-bootstrap-token\") pod \"machine-config-server-c9qdp\" (UID: \"437b5573-a342-4383-ba60-be0e3ccba839\") " pod="openshift-machine-config-operator/machine-config-server-c9qdp" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.716081 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7-config-volume\") pod \"collect-profiles-29490780-8q6q4\" (UID: \"7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-8q6q4" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.716212 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6clr\" (UniqueName: \"kubernetes.io/projected/37a77f41-5dbf-4842-9e77-83dc22b50f4a-kube-api-access-w6clr\") pod \"migrator-59844c95c7-qdmxz\" (UID: \"37a77f41-5dbf-4842-9e77-83dc22b50f4a\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qdmxz" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.716327 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8qsf\" (UniqueName: \"kubernetes.io/projected/2d37efbf-d18f-486b-9b43-bc4d181af4ca-kube-api-access-b8qsf\") pod \"marketplace-operator-79b997595-wvttb\" (UID: \"2d37efbf-d18f-486b-9b43-bc4d181af4ca\") " pod="openshift-marketplace/marketplace-operator-79b997595-wvttb" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.716435 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hc4dg\" (UniqueName: \"kubernetes.io/projected/abbeffe1-cfd5-4476-9a8e-2ab5b4869444-kube-api-access-hc4dg\") pod \"catalog-operator-68c6474976-nn46h\" (UID: \"abbeffe1-cfd5-4476-9a8e-2ab5b4869444\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nn46h" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.716642 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/c8657575-cd22-4ebc-ae9d-4174366985d3-csi-data-dir\") pod \"csi-hostpathplugin-vfm8t\" (UID: \"c8657575-cd22-4ebc-ae9d-4174366985d3\") " pod="hostpath-provisioner/csi-hostpathplugin-vfm8t" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.716884 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgzl5\" (UniqueName: \"kubernetes.io/projected/c8657575-cd22-4ebc-ae9d-4174366985d3-kube-api-access-fgzl5\") pod \"csi-hostpathplugin-vfm8t\" (UID: \"c8657575-cd22-4ebc-ae9d-4174366985d3\") " pod="hostpath-provisioner/csi-hostpathplugin-vfm8t" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.716847 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/c8657575-cd22-4ebc-ae9d-4174366985d3-csi-data-dir\") pod \"csi-hostpathplugin-vfm8t\" (UID: \"c8657575-cd22-4ebc-ae9d-4174366985d3\") " pod="hostpath-provisioner/csi-hostpathplugin-vfm8t" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.718041 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a4d83db5-776f-4e95-a6fa-b194344f9819-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-2sfhr\" (UID: \"a4d83db5-776f-4e95-a6fa-b194344f9819\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2sfhr" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.717094 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a4d83db5-776f-4e95-a6fa-b194344f9819-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-2sfhr\" (UID: \"a4d83db5-776f-4e95-a6fa-b194344f9819\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2sfhr" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.721436 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4z5z\" (UniqueName: \"kubernetes.io/projected/05d74105-0ecd-41ac-9001-8b21b0fd6ba4-kube-api-access-m4z5z\") pod \"openshift-controller-manager-operator-756b6f6bc6-l9nqd\" (UID: \"05d74105-0ecd-41ac-9001-8b21b0fd6ba4\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l9nqd" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.721588 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2d37efbf-d18f-486b-9b43-bc4d181af4ca-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wvttb\" (UID: \"2d37efbf-d18f-486b-9b43-bc4d181af4ca\") " pod="openshift-marketplace/marketplace-operator-79b997595-wvttb" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.721705 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/550752e4-a1d9-46f4-9118-9e9919b2fe6b-serving-cert\") pod \"service-ca-operator-777779d784-cjzsq\" (UID: \"550752e4-a1d9-46f4-9118-9e9919b2fe6b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cjzsq" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.721808 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ac10f013-cd1f-47e0-8f1c-5ff4e6e75784-config-volume\") pod \"dns-default-dgcqn\" (UID: \"ac10f013-cd1f-47e0-8f1c-5ff4e6e75784\") " pod="openshift-dns/dns-default-dgcqn" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.721962 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/c8657575-cd22-4ebc-ae9d-4174366985d3-mountpoint-dir\") pod \"csi-hostpathplugin-vfm8t\" (UID: \"c8657575-cd22-4ebc-ae9d-4174366985d3\") " pod="hostpath-provisioner/csi-hostpathplugin-vfm8t" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.722079 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/c8657575-cd22-4ebc-ae9d-4174366985d3-plugins-dir\") pod \"csi-hostpathplugin-vfm8t\" (UID: \"c8657575-cd22-4ebc-ae9d-4174366985d3\") " pod="hostpath-provisioner/csi-hostpathplugin-vfm8t" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.722175 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/c8657575-cd22-4ebc-ae9d-4174366985d3-mountpoint-dir\") pod \"csi-hostpathplugin-vfm8t\" (UID: \"c8657575-cd22-4ebc-ae9d-4174366985d3\") " pod="hostpath-provisioner/csi-hostpathplugin-vfm8t" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.722235 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2hpl\" (UniqueName: \"kubernetes.io/projected/a4d83db5-776f-4e95-a6fa-b194344f9819-kube-api-access-t2hpl\") pod \"machine-config-controller-84d6567774-2sfhr\" (UID: \"a4d83db5-776f-4e95-a6fa-b194344f9819\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2sfhr" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.722446 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/129a0b30-7132-4e3c-ab84-208cae7cb2f2-proxy-tls\") pod \"machine-config-operator-74547568cd-zzxln\" (UID: \"129a0b30-7132-4e3c-ab84-208cae7cb2f2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-zzxln" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.722350 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/c8657575-cd22-4ebc-ae9d-4174366985d3-plugins-dir\") pod \"csi-hostpathplugin-vfm8t\" (UID: \"c8657575-cd22-4ebc-ae9d-4174366985d3\") " pod="hostpath-provisioner/csi-hostpathplugin-vfm8t" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.724931 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.732880 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/73de6ef2-e139-4185-9f56-9db885734ffe-trusted-ca\") pod \"ingress-operator-5b745b69d9-58fcz\" (UID: \"73de6ef2-e139-4185-9f56-9db885734ffe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-58fcz" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.736905 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.756780 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.777133 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.796521 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.802291 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/85f05bd5-ff83-4d29-9531-ab3499088095-default-certificate\") pod \"router-default-5444994796-h9b2g\" (UID: \"85f05bd5-ff83-4d29-9531-ab3499088095\") " pod="openshift-ingress/router-default-5444994796-h9b2g" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.816581 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.827113 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/85f05bd5-ff83-4d29-9531-ab3499088095-stats-auth\") pod \"router-default-5444994796-h9b2g\" (UID: \"85f05bd5-ff83-4d29-9531-ab3499088095\") " pod="openshift-ingress/router-default-5444994796-h9b2g" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.836552 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.849456 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/85f05bd5-ff83-4d29-9531-ab3499088095-metrics-certs\") pod \"router-default-5444994796-h9b2g\" (UID: \"85f05bd5-ff83-4d29-9531-ab3499088095\") " pod="openshift-ingress/router-default-5444994796-h9b2g" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.856684 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.865023 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/85f05bd5-ff83-4d29-9531-ab3499088095-service-ca-bundle\") pod \"router-default-5444994796-h9b2g\" (UID: \"85f05bd5-ff83-4d29-9531-ab3499088095\") " pod="openshift-ingress/router-default-5444994796-h9b2g" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.876736 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.896810 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.916740 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.937375 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.949321 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa11789e-7a2a-4dbf-85ca-c20a9d64a1f4-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-6snv6\" (UID: \"fa11789e-7a2a-4dbf-85ca-c20a9d64a1f4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6snv6" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.956702 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.958190 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa11789e-7a2a-4dbf-85ca-c20a9d64a1f4-config\") pod \"kube-controller-manager-operator-78b949d7b-6snv6\" (UID: \"fa11789e-7a2a-4dbf-85ca-c20a9d64a1f4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6snv6" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.978979 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Jan 26 17:00:51 crc kubenswrapper[4856]: I0126 17:00:51.997323 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.016853 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.027780 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/42c0e428-821f-45a1-85a7-54ebdb81ef1c-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cl895\" (UID: \"42c0e428-821f-45a1-85a7-54ebdb81ef1c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cl895" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.037350 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.047331 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42c0e428-821f-45a1-85a7-54ebdb81ef1c-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cl895\" (UID: \"42c0e428-821f-45a1-85a7-54ebdb81ef1c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cl895" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.056679 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.077451 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.084269 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a4d83db5-776f-4e95-a6fa-b194344f9819-proxy-tls\") pod \"machine-config-controller-84d6567774-2sfhr\" (UID: \"a4d83db5-776f-4e95-a6fa-b194344f9819\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2sfhr" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.097199 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.101088 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/beb6f283-75cb-4184-b985-4e6c095feca1-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-ddghz\" (UID: \"beb6f283-75cb-4184-b985-4e6c095feca1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ddghz" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.117283 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.136827 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.146035 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/129a0b30-7132-4e3c-ab84-208cae7cb2f2-proxy-tls\") pod \"machine-config-operator-74547568cd-zzxln\" (UID: \"129a0b30-7132-4e3c-ab84-208cae7cb2f2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-zzxln" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.156705 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.163992 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/129a0b30-7132-4e3c-ab84-208cae7cb2f2-images\") pod \"machine-config-operator-74547568cd-zzxln\" (UID: \"129a0b30-7132-4e3c-ab84-208cae7cb2f2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-zzxln" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.176675 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.197265 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.204693 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/17a72e73-4d54-4a29-a85a-ecb1aff30d10-srv-cert\") pod \"olm-operator-6b444d44fb-k662z\" (UID: \"17a72e73-4d54-4a29-a85a-ecb1aff30d10\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k662z" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.217259 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.226061 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/17a72e73-4d54-4a29-a85a-ecb1aff30d10-profile-collector-cert\") pod \"olm-operator-6b444d44fb-k662z\" (UID: \"17a72e73-4d54-4a29-a85a-ecb1aff30d10\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k662z" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.226255 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7-secret-volume\") pod \"collect-profiles-29490780-8q6q4\" (UID: \"7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-8q6q4" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.228478 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/abbeffe1-cfd5-4476-9a8e-2ab5b4869444-profile-collector-cert\") pod \"catalog-operator-68c6474976-nn46h\" (UID: \"abbeffe1-cfd5-4476-9a8e-2ab5b4869444\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nn46h" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.237796 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.257400 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.276582 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.298671 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.305046 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/abbeffe1-cfd5-4476-9a8e-2ab5b4869444-srv-cert\") pod \"catalog-operator-68c6474976-nn46h\" (UID: \"abbeffe1-cfd5-4476-9a8e-2ab5b4869444\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nn46h" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.316900 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.324701 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/004316da-16cd-49ab-b14d-282c28da6fad-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-8m4l6\" (UID: \"004316da-16cd-49ab-b14d-282c28da6fad\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8m4l6" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.337743 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.341048 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f9b8f57e-00b9-4355-ace2-0319d320d208-apiservice-cert\") pod \"packageserver-d55dfcdfc-mr7cp\" (UID: \"f9b8f57e-00b9-4355-ace2-0319d320d208\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mr7cp" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.347504 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f9b8f57e-00b9-4355-ace2-0319d320d208-webhook-cert\") pod \"packageserver-d55dfcdfc-mr7cp\" (UID: \"f9b8f57e-00b9-4355-ace2-0319d320d208\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mr7cp" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.356405 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.364856 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05d74105-0ecd-41ac-9001-8b21b0fd6ba4-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-l9nqd\" (UID: \"05d74105-0ecd-41ac-9001-8b21b0fd6ba4\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l9nqd" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.376793 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.394255 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.394350 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.394354 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.394772 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.396223 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.417002 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.436372 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.450738 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/05d74105-0ecd-41ac-9001-8b21b0fd6ba4-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-l9nqd\" (UID: \"05d74105-0ecd-41ac-9001-8b21b0fd6ba4\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l9nqd" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.456691 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.477346 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.494760 4856 request.go:700] Waited for 1.017759772s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-service-ca/secrets?fieldSelector=metadata.name%3Dsigning-key&limit=500&resourceVersion=0 Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.496663 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.501067 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/033cb12f-278f-431a-8104-519db9a3152f-signing-key\") pod \"service-ca-9c57cc56f-gz7kg\" (UID: \"033cb12f-278f-431a-8104-519db9a3152f\") " pod="openshift-service-ca/service-ca-9c57cc56f-gz7kg" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.516987 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.524362 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/033cb12f-278f-431a-8104-519db9a3152f-signing-cabundle\") pod \"service-ca-9c57cc56f-gz7kg\" (UID: \"033cb12f-278f-431a-8104-519db9a3152f\") " pod="openshift-service-ca/service-ca-9c57cc56f-gz7kg" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.537613 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.577439 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.596659 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Jan 26 17:00:52 crc kubenswrapper[4856]: E0126 17:00:52.598095 4856 secret.go:188] Couldn't get secret openshift-kube-apiserver-operator/kube-apiserver-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Jan 26 17:00:52 crc kubenswrapper[4856]: E0126 17:00:52.598214 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ddc2e6b7-5582-4579-bf2c-ed165b74c91a-serving-cert podName:ddc2e6b7-5582-4579-bf2c-ed165b74c91a nodeName:}" failed. No retries permitted until 2026-01-26 17:00:53.098174872 +0000 UTC m=+149.051428853 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/ddc2e6b7-5582-4579-bf2c-ed165b74c91a-serving-cert") pod "kube-apiserver-operator-766d6c64bb-vmwvg" (UID: "ddc2e6b7-5582-4579-bf2c-ed165b74c91a") : failed to sync secret cache: timed out waiting for the condition Jan 26 17:00:52 crc kubenswrapper[4856]: E0126 17:00:52.599155 4856 configmap.go:193] Couldn't get configMap openshift-kube-apiserver-operator/kube-apiserver-operator-config: failed to sync configmap cache: timed out waiting for the condition Jan 26 17:00:52 crc kubenswrapper[4856]: E0126 17:00:52.599193 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/ddc2e6b7-5582-4579-bf2c-ed165b74c91a-config podName:ddc2e6b7-5582-4579-bf2c-ed165b74c91a nodeName:}" failed. No retries permitted until 2026-01-26 17:00:53.099183081 +0000 UTC m=+149.052437062 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/ddc2e6b7-5582-4579-bf2c-ed165b74c91a-config") pod "kube-apiserver-operator-766d6c64bb-vmwvg" (UID: "ddc2e6b7-5582-4579-bf2c-ed165b74c91a") : failed to sync configmap cache: timed out waiting for the condition Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.632587 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.636859 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.648199 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2d37efbf-d18f-486b-9b43-bc4d181af4ca-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wvttb\" (UID: \"2d37efbf-d18f-486b-9b43-bc4d181af4ca\") " pod="openshift-marketplace/marketplace-operator-79b997595-wvttb" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.664313 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.673996 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2d37efbf-d18f-486b-9b43-bc4d181af4ca-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wvttb\" (UID: \"2d37efbf-d18f-486b-9b43-bc4d181af4ca\") " pod="openshift-marketplace/marketplace-operator-79b997595-wvttb" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.677282 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.696624 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.707641 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7-config-volume\") pod \"collect-profiles-29490780-8q6q4\" (UID: \"7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-8q6q4" Jan 26 17:00:52 crc kubenswrapper[4856]: E0126 17:00:52.711584 4856 secret.go:188] Couldn't get secret openshift-dns/dns-default-metrics-tls: failed to sync secret cache: timed out waiting for the condition Jan 26 17:00:52 crc kubenswrapper[4856]: E0126 17:00:52.711615 4856 secret.go:188] Couldn't get secret openshift-machine-api/control-plane-machine-set-operator-tls: failed to sync secret cache: timed out waiting for the condition Jan 26 17:00:52 crc kubenswrapper[4856]: E0126 17:00:52.711583 4856 configmap.go:193] Couldn't get configMap openshift-service-ca-operator/service-ca-operator-config: failed to sync configmap cache: timed out waiting for the condition Jan 26 17:00:52 crc kubenswrapper[4856]: E0126 17:00:52.711681 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a1546392-4a69-4b12-8d7e-97450b73b7ca-control-plane-machine-set-operator-tls podName:a1546392-4a69-4b12-8d7e-97450b73b7ca nodeName:}" failed. No retries permitted until 2026-01-26 17:00:53.211659938 +0000 UTC m=+149.164913919 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "control-plane-machine-set-operator-tls" (UniqueName: "kubernetes.io/secret/a1546392-4a69-4b12-8d7e-97450b73b7ca-control-plane-machine-set-operator-tls") pod "control-plane-machine-set-operator-78cbb6b69f-rrhjv" (UID: "a1546392-4a69-4b12-8d7e-97450b73b7ca") : failed to sync secret cache: timed out waiting for the condition Jan 26 17:00:52 crc kubenswrapper[4856]: E0126 17:00:52.711698 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ac10f013-cd1f-47e0-8f1c-5ff4e6e75784-metrics-tls podName:ac10f013-cd1f-47e0-8f1c-5ff4e6e75784 nodeName:}" failed. No retries permitted until 2026-01-26 17:00:53.211691179 +0000 UTC m=+149.164945160 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/ac10f013-cd1f-47e0-8f1c-5ff4e6e75784-metrics-tls") pod "dns-default-dgcqn" (UID: "ac10f013-cd1f-47e0-8f1c-5ff4e6e75784") : failed to sync secret cache: timed out waiting for the condition Jan 26 17:00:52 crc kubenswrapper[4856]: E0126 17:00:52.711765 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/550752e4-a1d9-46f4-9118-9e9919b2fe6b-config podName:550752e4-a1d9-46f4-9118-9e9919b2fe6b nodeName:}" failed. No retries permitted until 2026-01-26 17:00:53.21172791 +0000 UTC m=+149.164981961 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/550752e4-a1d9-46f4-9118-9e9919b2fe6b-config") pod "service-ca-operator-777779d784-cjzsq" (UID: "550752e4-a1d9-46f4-9118-9e9919b2fe6b") : failed to sync configmap cache: timed out waiting for the condition Jan 26 17:00:52 crc kubenswrapper[4856]: E0126 17:00:52.712817 4856 secret.go:188] Couldn't get secret openshift-machine-config-operator/machine-config-server-tls: failed to sync secret cache: timed out waiting for the condition Jan 26 17:00:52 crc kubenswrapper[4856]: E0126 17:00:52.712875 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/437b5573-a342-4383-ba60-be0e3ccba839-certs podName:437b5573-a342-4383-ba60-be0e3ccba839 nodeName:}" failed. No retries permitted until 2026-01-26 17:00:53.212858843 +0000 UTC m=+149.166112885 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "certs" (UniqueName: "kubernetes.io/secret/437b5573-a342-4383-ba60-be0e3ccba839-certs") pod "machine-config-server-c9qdp" (UID: "437b5573-a342-4383-ba60-be0e3ccba839") : failed to sync secret cache: timed out waiting for the condition Jan 26 17:00:52 crc kubenswrapper[4856]: E0126 17:00:52.716106 4856 secret.go:188] Couldn't get secret openshift-machine-config-operator/node-bootstrapper-token: failed to sync secret cache: timed out waiting for the condition Jan 26 17:00:52 crc kubenswrapper[4856]: E0126 17:00:52.716218 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/437b5573-a342-4383-ba60-be0e3ccba839-node-bootstrap-token podName:437b5573-a342-4383-ba60-be0e3ccba839 nodeName:}" failed. No retries permitted until 2026-01-26 17:00:53.216203682 +0000 UTC m=+149.169457663 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-bootstrap-token" (UniqueName: "kubernetes.io/secret/437b5573-a342-4383-ba60-be0e3ccba839-node-bootstrap-token") pod "machine-config-server-c9qdp" (UID: "437b5573-a342-4383-ba60-be0e3ccba839") : failed to sync secret cache: timed out waiting for the condition Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.716995 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Jan 26 17:00:52 crc kubenswrapper[4856]: E0126 17:00:52.722595 4856 configmap.go:193] Couldn't get configMap openshift-dns/dns-default: failed to sync configmap cache: timed out waiting for the condition Jan 26 17:00:52 crc kubenswrapper[4856]: E0126 17:00:52.722662 4856 secret.go:188] Couldn't get secret openshift-service-ca-operator/serving-cert: failed to sync secret cache: timed out waiting for the condition Jan 26 17:00:52 crc kubenswrapper[4856]: E0126 17:00:52.722686 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/ac10f013-cd1f-47e0-8f1c-5ff4e6e75784-config-volume podName:ac10f013-cd1f-47e0-8f1c-5ff4e6e75784 nodeName:}" failed. No retries permitted until 2026-01-26 17:00:53.222666523 +0000 UTC m=+149.175920504 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/ac10f013-cd1f-47e0-8f1c-5ff4e6e75784-config-volume") pod "dns-default-dgcqn" (UID: "ac10f013-cd1f-47e0-8f1c-5ff4e6e75784") : failed to sync configmap cache: timed out waiting for the condition Jan 26 17:00:52 crc kubenswrapper[4856]: E0126 17:00:52.722719 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/550752e4-a1d9-46f4-9118-9e9919b2fe6b-serving-cert podName:550752e4-a1d9-46f4-9118-9e9919b2fe6b nodeName:}" failed. No retries permitted until 2026-01-26 17:00:53.222705144 +0000 UTC m=+149.175959125 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/550752e4-a1d9-46f4-9118-9e9919b2fe6b-serving-cert") pod "service-ca-operator-777779d784-cjzsq" (UID: "550752e4-a1d9-46f4-9118-9e9919b2fe6b") : failed to sync secret cache: timed out waiting for the condition Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.736814 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.756729 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.776943 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.796275 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.816723 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.836689 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.856472 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.877053 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.897900 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.917570 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.937512 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.957286 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Jan 26 17:00:52 crc kubenswrapper[4856]: I0126 17:00:52.978865 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.013329 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tmd5\" (UniqueName: \"kubernetes.io/projected/2ba3cf6a-a6be-4108-a155-c8bb530aa037-kube-api-access-6tmd5\") pod \"openshift-config-operator-7777fb866f-5bjl7\" (UID: \"2ba3cf6a-a6be-4108-a155-c8bb530aa037\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bjl7" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.017511 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.037914 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.057020 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.076959 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.096409 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.117584 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.137482 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.149949 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddc2e6b7-5582-4579-bf2c-ed165b74c91a-config\") pod \"kube-apiserver-operator-766d6c64bb-vmwvg\" (UID: \"ddc2e6b7-5582-4579-bf2c-ed165b74c91a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vmwvg" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.150362 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ddc2e6b7-5582-4579-bf2c-ed165b74c91a-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-vmwvg\" (UID: \"ddc2e6b7-5582-4579-bf2c-ed165b74c91a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vmwvg" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.150988 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddc2e6b7-5582-4579-bf2c-ed165b74c91a-config\") pod \"kube-apiserver-operator-766d6c64bb-vmwvg\" (UID: \"ddc2e6b7-5582-4579-bf2c-ed165b74c91a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vmwvg" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.153582 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ddc2e6b7-5582-4579-bf2c-ed165b74c91a-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-vmwvg\" (UID: \"ddc2e6b7-5582-4579-bf2c-ed165b74c91a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vmwvg" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.157400 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.177045 4856 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.196929 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.217446 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.237143 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.252108 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/437b5573-a342-4383-ba60-be0e3ccba839-node-bootstrap-token\") pod \"machine-config-server-c9qdp\" (UID: \"437b5573-a342-4383-ba60-be0e3ccba839\") " pod="openshift-machine-config-operator/machine-config-server-c9qdp" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.252264 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/550752e4-a1d9-46f4-9118-9e9919b2fe6b-serving-cert\") pod \"service-ca-operator-777779d784-cjzsq\" (UID: \"550752e4-a1d9-46f4-9118-9e9919b2fe6b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cjzsq" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.252307 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ac10f013-cd1f-47e0-8f1c-5ff4e6e75784-config-volume\") pod \"dns-default-dgcqn\" (UID: \"ac10f013-cd1f-47e0-8f1c-5ff4e6e75784\") " pod="openshift-dns/dns-default-dgcqn" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.252452 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ac10f013-cd1f-47e0-8f1c-5ff4e6e75784-metrics-tls\") pod \"dns-default-dgcqn\" (UID: \"ac10f013-cd1f-47e0-8f1c-5ff4e6e75784\") " pod="openshift-dns/dns-default-dgcqn" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.252488 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/550752e4-a1d9-46f4-9118-9e9919b2fe6b-config\") pod \"service-ca-operator-777779d784-cjzsq\" (UID: \"550752e4-a1d9-46f4-9118-9e9919b2fe6b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cjzsq" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.252746 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a1546392-4a69-4b12-8d7e-97450b73b7ca-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-rrhjv\" (UID: \"a1546392-4a69-4b12-8d7e-97450b73b7ca\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rrhjv" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.252869 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/437b5573-a342-4383-ba60-be0e3ccba839-certs\") pod \"machine-config-server-c9qdp\" (UID: \"437b5573-a342-4383-ba60-be0e3ccba839\") " pod="openshift-machine-config-operator/machine-config-server-c9qdp" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.253242 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/550752e4-a1d9-46f4-9118-9e9919b2fe6b-config\") pod \"service-ca-operator-777779d784-cjzsq\" (UID: \"550752e4-a1d9-46f4-9118-9e9919b2fe6b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cjzsq" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.255327 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/550752e4-a1d9-46f4-9118-9e9919b2fe6b-serving-cert\") pod \"service-ca-operator-777779d784-cjzsq\" (UID: \"550752e4-a1d9-46f4-9118-9e9919b2fe6b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cjzsq" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.255360 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a1546392-4a69-4b12-8d7e-97450b73b7ca-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-rrhjv\" (UID: \"a1546392-4a69-4b12-8d7e-97450b73b7ca\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rrhjv" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.256982 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.260698 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bjl7" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.266740 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/437b5573-a342-4383-ba60-be0e3ccba839-certs\") pod \"machine-config-server-c9qdp\" (UID: \"437b5573-a342-4383-ba60-be0e3ccba839\") " pod="openshift-machine-config-operator/machine-config-server-c9qdp" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.278583 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.286999 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/437b5573-a342-4383-ba60-be0e3ccba839-node-bootstrap-token\") pod \"machine-config-server-c9qdp\" (UID: \"437b5573-a342-4383-ba60-be0e3ccba839\") " pod="openshift-machine-config-operator/machine-config-server-c9qdp" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.296742 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.317110 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.327228 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ac10f013-cd1f-47e0-8f1c-5ff4e6e75784-metrics-tls\") pod \"dns-default-dgcqn\" (UID: \"ac10f013-cd1f-47e0-8f1c-5ff4e6e75784\") " pod="openshift-dns/dns-default-dgcqn" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.336282 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.343374 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ac10f013-cd1f-47e0-8f1c-5ff4e6e75784-config-volume\") pod \"dns-default-dgcqn\" (UID: \"ac10f013-cd1f-47e0-8f1c-5ff4e6e75784\") " pod="openshift-dns/dns-default-dgcqn" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.394026 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdh5l\" (UniqueName: \"kubernetes.io/projected/149e3000-35d7-47bd-83f0-00ab5e0736c2-kube-api-access-mdh5l\") pod \"kube-storage-version-migrator-operator-b67b599dd-q7j7b\" (UID: \"149e3000-35d7-47bd-83f0-00ab5e0736c2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-q7j7b" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.412151 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mr8gn\" (UniqueName: \"kubernetes.io/projected/1afc0f4c-e02d-4a70-aaba-e761e8c04eee-kube-api-access-mr8gn\") pod \"controller-manager-879f6c89f-lndnt\" (UID: \"1afc0f4c-e02d-4a70-aaba-e761e8c04eee\") " pod="openshift-controller-manager/controller-manager-879f6c89f-lndnt" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.433594 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lg98g\" (UniqueName: \"kubernetes.io/projected/a6d331bd-2db3-4319-9f5c-db56d408d9e3-kube-api-access-lg98g\") pod \"apiserver-76f77b778f-6rlxp\" (UID: \"a6d331bd-2db3-4319-9f5c-db56d408d9e3\") " pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.471672 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf1f11c8-17b8-49b7-b12d-92891f478222-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-7p5jt\" (UID: \"bf1f11c8-17b8-49b7-b12d-92891f478222\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7p5jt" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.502158 4856 request.go:700] Waited for 1.90475386s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-apiserver-operator/serviceaccounts/openshift-apiserver-operator/token Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.503284 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpfwk\" (UniqueName: \"kubernetes.io/projected/5fe6baed-ab97-4d8a-8be2-6f00f9698136-kube-api-access-hpfwk\") pod \"route-controller-manager-6576b87f9c-fpqvc\" (UID: \"5fe6baed-ab97-4d8a-8be2-6f00f9698136\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpqvc" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.504773 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkngl\" (UniqueName: \"kubernetes.io/projected/5c244eff-aada-44f3-b250-96878a3400c4-kube-api-access-nkngl\") pod \"etcd-operator-b45778765-27vjc\" (UID: \"5c244eff-aada-44f3-b250-96878a3400c4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-27vjc" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.518271 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkl7p\" (UniqueName: \"kubernetes.io/projected/359660cd-b412-4640-bedf-993e976e7b3c-kube-api-access-rkl7p\") pod \"openshift-apiserver-operator-796bbdcf4f-88lkr\" (UID: \"359660cd-b412-4640-bedf-993e976e7b3c\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-88lkr" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.532432 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/42c0e428-821f-45a1-85a7-54ebdb81ef1c-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-cl895\" (UID: \"42c0e428-821f-45a1-85a7-54ebdb81ef1c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cl895" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.552457 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vq4z\" (UniqueName: \"kubernetes.io/projected/a7e5d16a-d45d-4b40-93bf-bfaa6be2d1c0-kube-api-access-7vq4z\") pod \"authentication-operator-69f744f599-jdjcq\" (UID: \"a7e5d16a-d45d-4b40-93bf-bfaa6be2d1c0\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jdjcq" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.571337 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6fxv\" (UniqueName: \"kubernetes.io/projected/f9b8f57e-00b9-4355-ace2-0319d320d208-kube-api-access-d6fxv\") pod \"packageserver-d55dfcdfc-mr7cp\" (UID: \"f9b8f57e-00b9-4355-ace2-0319d320d208\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mr7cp" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.574001 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-5bjl7"] Jan 26 17:00:53 crc kubenswrapper[4856]: W0126 17:00:53.582509 4856 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2ba3cf6a_a6be_4108_a155_c8bb530aa037.slice/crio-f9c4b2e98acee19d774149a7d740d7163dacd0dcb0f25054caba9129d5ee7274 WatchSource:0}: Error finding container f9c4b2e98acee19d774149a7d740d7163dacd0dcb0f25054caba9129d5ee7274: Status 404 returned error can't find the container with id f9c4b2e98acee19d774149a7d740d7163dacd0dcb0f25054caba9129d5ee7274 Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.595470 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kf2b2\" (UniqueName: \"kubernetes.io/projected/69008ed1-f3e5-400d-852f-adbcd94199f6-kube-api-access-kf2b2\") pod \"oauth-openshift-558db77b4-cb8nk\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.612871 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvkhn\" (UniqueName: \"kubernetes.io/projected/b28404ed-2e71-4b3f-9140-35ee89dbc8f2-kube-api-access-dvkhn\") pod \"console-f9d7485db-6qgnn\" (UID: \"b28404ed-2e71-4b3f-9140-35ee89dbc8f2\") " pod="openshift-console/console-f9d7485db-6qgnn" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.616795 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-88lkr" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.637362 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2274g\" (UniqueName: \"kubernetes.io/projected/54a246a2-f674-4735-b295-b56699ece95b-kube-api-access-2274g\") pod \"machine-approver-56656f9798-962cr\" (UID: \"54a246a2-f674-4735-b295-b56699ece95b\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-962cr" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.639981 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-6qgnn" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.652007 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-jdjcq" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.653069 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fa11789e-7a2a-4dbf-85ca-c20a9d64a1f4-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-6snv6\" (UID: \"fa11789e-7a2a-4dbf-85ca-c20a9d64a1f4\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6snv6" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.669974 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-27vjc" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.675172 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5scg\" (UniqueName: \"kubernetes.io/projected/033cb12f-278f-431a-8104-519db9a3152f-kube-api-access-s5scg\") pod \"service-ca-9c57cc56f-gz7kg\" (UID: \"033cb12f-278f-431a-8104-519db9a3152f\") " pod="openshift-service-ca/service-ca-9c57cc56f-gz7kg" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.677901 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-q7j7b" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.695338 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.696909 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2wtm\" (UniqueName: \"kubernetes.io/projected/0c1af7db-aa80-4cb0-a9cb-5afdf677f28c-kube-api-access-v2wtm\") pod \"cluster-samples-operator-665b6dd947-4w5bf\" (UID: \"0c1af7db-aa80-4cb0-a9cb-5afdf677f28c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4w5bf" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.706005 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-lndnt" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.711475 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qcw5\" (UniqueName: \"kubernetes.io/projected/bf1f11c8-17b8-49b7-b12d-92891f478222-kube-api-access-9qcw5\") pod \"cluster-image-registry-operator-dc59b4c8b-7p5jt\" (UID: \"bf1f11c8-17b8-49b7-b12d-92891f478222\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7p5jt" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.736144 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgb6k\" (UniqueName: \"kubernetes.io/projected/94291fa4-24a5-499e-8143-89c8784d9284-kube-api-access-hgb6k\") pod \"downloads-7954f5f757-7l927\" (UID: \"94291fa4-24a5-499e-8143-89c8784d9284\") " pod="openshift-console/downloads-7954f5f757-7l927" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.760076 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ddc2e6b7-5582-4579-bf2c-ed165b74c91a-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-vmwvg\" (UID: \"ddc2e6b7-5582-4579-bf2c-ed165b74c91a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vmwvg" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.765131 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6snv6" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.783205 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2wnd\" (UniqueName: \"kubernetes.io/projected/831dc87e-8e14-43d3-a36e-dc7679041ae5-kube-api-access-d2wnd\") pod \"console-operator-58897d9998-4pbj2\" (UID: \"831dc87e-8e14-43d3-a36e-dc7679041ae5\") " pod="openshift-console-operator/console-operator-58897d9998-4pbj2" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.785893 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpqvc" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.800304 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cl895" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.803934 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqhmb\" (UniqueName: \"kubernetes.io/projected/81c2f96b-55e0-483b-b72c-df7e156e9218-kube-api-access-rqhmb\") pod \"apiserver-7bbb656c7d-6cghs\" (UID: \"81c2f96b-55e0-483b-b72c-df7e156e9218\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6cghs" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.835956 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-88lkr"] Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.838090 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4w5bf" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.838914 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjhr5\" (UniqueName: \"kubernetes.io/projected/beb6f283-75cb-4184-b985-4e6c095feca1-kube-api-access-mjhr5\") pod \"multus-admission-controller-857f4d67dd-ddghz\" (UID: \"beb6f283-75cb-4184-b985-4e6c095feca1\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-ddghz" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.844126 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7lk4\" (UniqueName: \"kubernetes.io/projected/85f05bd5-ff83-4d29-9531-ab3499088095-kube-api-access-x7lk4\") pod \"router-default-5444994796-h9b2g\" (UID: \"85f05bd5-ff83-4d29-9531-ab3499088095\") " pod="openshift-ingress/router-default-5444994796-h9b2g" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.848772 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-7l927" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.851848 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mr7cp" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.861735 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c475g\" (UniqueName: \"kubernetes.io/projected/77a97acb-2908-48fb-8bcd-0647f3e90160-kube-api-access-c475g\") pod \"machine-api-operator-5694c8668f-7xb2b\" (UID: \"77a97acb-2908-48fb-8bcd-0647f3e90160\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7xb2b" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.867297 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-gz7kg" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.873280 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nr9jp\" (UniqueName: \"kubernetes.io/projected/004316da-16cd-49ab-b14d-282c28da6fad-kube-api-access-nr9jp\") pod \"package-server-manager-789f6589d5-8m4l6\" (UID: \"004316da-16cd-49ab-b14d-282c28da6fad\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8m4l6" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.873658 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.873765 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-6qgnn"] Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.885897 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7p5jt" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.894477 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-4pbj2" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.895234 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/73de6ef2-e139-4185-9f56-9db885734ffe-bound-sa-token\") pod \"ingress-operator-5b745b69d9-58fcz\" (UID: \"73de6ef2-e139-4185-9f56-9db885734ffe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-58fcz" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.910693 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6v8bp\" (UniqueName: \"kubernetes.io/projected/ac10f013-cd1f-47e0-8f1c-5ff4e6e75784-kube-api-access-6v8bp\") pod \"dns-default-dgcqn\" (UID: \"ac10f013-cd1f-47e0-8f1c-5ff4e6e75784\") " pod="openshift-dns/dns-default-dgcqn" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.911293 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vmwvg" Jan 26 17:00:53 crc kubenswrapper[4856]: W0126 17:00:53.918238 4856 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod359660cd_b412_4640_bedf_993e976e7b3c.slice/crio-67342ac4fc36ecabd1a17f918aa59bd99af37bddc54a8e3deb2760014c345a08 WatchSource:0}: Error finding container 67342ac4fc36ecabd1a17f918aa59bd99af37bddc54a8e3deb2760014c345a08: Status 404 returned error can't find the container with id 67342ac4fc36ecabd1a17f918aa59bd99af37bddc54a8e3deb2760014c345a08 Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.924386 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-962cr" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.946411 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zchmc\" (UniqueName: \"kubernetes.io/projected/550752e4-a1d9-46f4-9118-9e9919b2fe6b-kube-api-access-zchmc\") pod \"service-ca-operator-777779d784-cjzsq\" (UID: \"550752e4-a1d9-46f4-9118-9e9919b2fe6b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-cjzsq" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.956226 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vxn2\" (UniqueName: \"kubernetes.io/projected/437b5573-a342-4383-ba60-be0e3ccba839-kube-api-access-9vxn2\") pod \"machine-config-server-c9qdp\" (UID: \"437b5573-a342-4383-ba60-be0e3ccba839\") " pod="openshift-machine-config-operator/machine-config-server-c9qdp" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.960168 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-7xb2b" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.972347 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-c9qdp" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.979880 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-dgcqn" Jan 26 17:00:53 crc kubenswrapper[4856]: I0126 17:00:53.984020 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kkkd\" (UniqueName: \"kubernetes.io/projected/129a0b30-7132-4e3c-ab84-208cae7cb2f2-kube-api-access-6kkkd\") pod \"machine-config-operator-74547568cd-zzxln\" (UID: \"129a0b30-7132-4e3c-ab84-208cae7cb2f2\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-zzxln" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.006212 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdxxq\" (UniqueName: \"kubernetes.io/projected/73de6ef2-e139-4185-9f56-9db885734ffe-kube-api-access-hdxxq\") pod \"ingress-operator-5b745b69d9-58fcz\" (UID: \"73de6ef2-e139-4185-9f56-9db885734ffe\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-58fcz" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.034627 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6cghs" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.039731 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxgfj\" (UniqueName: \"kubernetes.io/projected/7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7-kube-api-access-kxgfj\") pod \"collect-profiles-29490780-8q6q4\" (UID: \"7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-8q6q4" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.053814 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-h9b2g" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.059944 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwmjs\" (UniqueName: \"kubernetes.io/projected/a1546392-4a69-4b12-8d7e-97450b73b7ca-kube-api-access-pwmjs\") pod \"control-plane-machine-set-operator-78cbb6b69f-rrhjv\" (UID: \"a1546392-4a69-4b12-8d7e-97450b73b7ca\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rrhjv" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.064562 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wxzv\" (UniqueName: \"kubernetes.io/projected/17a72e73-4d54-4a29-a85a-ecb1aff30d10-kube-api-access-9wxzv\") pod \"olm-operator-6b444d44fb-k662z\" (UID: \"17a72e73-4d54-4a29-a85a-ecb1aff30d10\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k662z" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.079023 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6clr\" (UniqueName: \"kubernetes.io/projected/37a77f41-5dbf-4842-9e77-83dc22b50f4a-kube-api-access-w6clr\") pod \"migrator-59844c95c7-qdmxz\" (UID: \"37a77f41-5dbf-4842-9e77-83dc22b50f4a\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qdmxz" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.103594 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-ddghz" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.104683 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8qsf\" (UniqueName: \"kubernetes.io/projected/2d37efbf-d18f-486b-9b43-bc4d181af4ca-kube-api-access-b8qsf\") pod \"marketplace-operator-79b997595-wvttb\" (UID: \"2d37efbf-d18f-486b-9b43-bc4d181af4ca\") " pod="openshift-marketplace/marketplace-operator-79b997595-wvttb" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.106917 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bjl7" event={"ID":"2ba3cf6a-a6be-4108-a155-c8bb530aa037","Type":"ContainerStarted","Data":"d89a3287b371ef0cb23bab1eb475eccd1999b48798e71ba74c19f952107aef34"} Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.107168 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bjl7" event={"ID":"2ba3cf6a-a6be-4108-a155-c8bb530aa037","Type":"ContainerStarted","Data":"f9c4b2e98acee19d774149a7d740d7163dacd0dcb0f25054caba9129d5ee7274"} Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.109125 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-6qgnn" event={"ID":"b28404ed-2e71-4b3f-9140-35ee89dbc8f2","Type":"ContainerStarted","Data":"8d60fe83a3b8c25a6706fa61b15688fd93c7ae27849eb62b7e61218c7bdddb31"} Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.110414 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-zzxln" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.112374 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-88lkr" event={"ID":"359660cd-b412-4640-bedf-993e976e7b3c","Type":"ContainerStarted","Data":"67342ac4fc36ecabd1a17f918aa59bd99af37bddc54a8e3deb2760014c345a08"} Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.123683 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hc4dg\" (UniqueName: \"kubernetes.io/projected/abbeffe1-cfd5-4476-9a8e-2ab5b4869444-kube-api-access-hc4dg\") pod \"catalog-operator-68c6474976-nn46h\" (UID: \"abbeffe1-cfd5-4476-9a8e-2ab5b4869444\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nn46h" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.132331 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k662z" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.137776 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nn46h" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.139263 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgzl5\" (UniqueName: \"kubernetes.io/projected/c8657575-cd22-4ebc-ae9d-4174366985d3-kube-api-access-fgzl5\") pod \"csi-hostpathplugin-vfm8t\" (UID: \"c8657575-cd22-4ebc-ae9d-4174366985d3\") " pod="hostpath-provisioner/csi-hostpathplugin-vfm8t" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.142867 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8m4l6" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.163423 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4z5z\" (UniqueName: \"kubernetes.io/projected/05d74105-0ecd-41ac-9001-8b21b0fd6ba4-kube-api-access-m4z5z\") pod \"openshift-controller-manager-operator-756b6f6bc6-l9nqd\" (UID: \"05d74105-0ecd-41ac-9001-8b21b0fd6ba4\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l9nqd" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.173855 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wvttb" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.180295 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.184004 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-jdjcq"] Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.184080 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-q7j7b"] Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.184093 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-8q6q4" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.187263 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2hpl\" (UniqueName: \"kubernetes.io/projected/a4d83db5-776f-4e95-a6fa-b194344f9819-kube-api-access-t2hpl\") pod \"machine-config-controller-84d6567774-2sfhr\" (UID: \"a4d83db5-776f-4e95-a6fa-b194344f9819\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2sfhr" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.192346 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rrhjv" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.200600 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.202022 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-cjzsq" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.218166 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.239004 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.258175 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.260684 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-vfm8t" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.266426 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-6rlxp"] Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.277860 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.285519 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qdmxz" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.287494 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lndnt"] Jan 26 17:00:54 crc kubenswrapper[4856]: W0126 17:00:54.288312 4856 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda7e5d16a_d45d_4b40_93bf_bfaa6be2d1c0.slice/crio-ef1cb2622920f923afb73fc12eda34877038d999d31cb4114611a29f8ca1fbd5 WatchSource:0}: Error finding container ef1cb2622920f923afb73fc12eda34877038d999d31cb4114611a29f8ca1fbd5: Status 404 returned error can't find the container with id ef1cb2622920f923afb73fc12eda34877038d999d31cb4114611a29f8ca1fbd5 Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.294118 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-58fcz" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.310543 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-27vjc"] Jan 26 17:00:54 crc kubenswrapper[4856]: W0126 17:00:54.345946 4856 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda6d331bd_2db3_4319_9f5c_db56d408d9e3.slice/crio-bae2856860ccbfbac0017e7da1d4f30234d58daef19aaf9d48d7644f9aac2b38 WatchSource:0}: Error finding container bae2856860ccbfbac0017e7da1d4f30234d58daef19aaf9d48d7644f9aac2b38: Status 404 returned error can't find the container with id bae2856860ccbfbac0017e7da1d4f30234d58daef19aaf9d48d7644f9aac2b38 Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.395917 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/113d2266-0e67-4e79-8a17-1a78cb9a13d5-metrics-tls\") pod \"dns-operator-744455d44c-z7cgq\" (UID: \"113d2266-0e67-4e79-8a17-1a78cb9a13d5\") " pod="openshift-dns-operator/dns-operator-744455d44c-z7cgq" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.396009 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cfa40861-cc08-4145-a185-6a3fb07eaabe-trusted-ca\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.396071 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tf448\" (UniqueName: \"kubernetes.io/projected/cfa40861-cc08-4145-a185-6a3fb07eaabe-kube-api-access-tf448\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.396245 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9tb9\" (UniqueName: \"kubernetes.io/projected/cb9fb12b-3eb8-4e94-a8cf-9eaf4703a850-kube-api-access-s9tb9\") pod \"ingress-canary-fbsj7\" (UID: \"cb9fb12b-3eb8-4e94-a8cf-9eaf4703a850\") " pod="openshift-ingress-canary/ingress-canary-fbsj7" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.396334 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5q847\" (UniqueName: \"kubernetes.io/projected/113d2266-0e67-4e79-8a17-1a78cb9a13d5-kube-api-access-5q847\") pod \"dns-operator-744455d44c-z7cgq\" (UID: \"113d2266-0e67-4e79-8a17-1a78cb9a13d5\") " pod="openshift-dns-operator/dns-operator-744455d44c-z7cgq" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.396441 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cfa40861-cc08-4145-a185-6a3fb07eaabe-installation-pull-secrets\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.396562 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.396664 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cb9fb12b-3eb8-4e94-a8cf-9eaf4703a850-cert\") pod \"ingress-canary-fbsj7\" (UID: \"cb9fb12b-3eb8-4e94-a8cf-9eaf4703a850\") " pod="openshift-ingress-canary/ingress-canary-fbsj7" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.396726 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cfa40861-cc08-4145-a185-6a3fb07eaabe-registry-certificates\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.396946 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cfa40861-cc08-4145-a185-6a3fb07eaabe-registry-tls\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.396974 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cfa40861-cc08-4145-a185-6a3fb07eaabe-ca-trust-extracted\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.397007 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cfa40861-cc08-4145-a185-6a3fb07eaabe-bound-sa-token\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:54 crc kubenswrapper[4856]: E0126 17:00:54.399329 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:00:54.899306536 +0000 UTC m=+150.852560567 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.403338 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2sfhr" Jan 26 17:00:54 crc kubenswrapper[4856]: W0126 17:00:54.407463 4856 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1afc0f4c_e02d_4a70_aaba_e761e8c04eee.slice/crio-96859d6a59b58c9df792a590deef50eb0ee923d03cb16fdc72abe3d18e466eaa WatchSource:0}: Error finding container 96859d6a59b58c9df792a590deef50eb0ee923d03cb16fdc72abe3d18e466eaa: Status 404 returned error can't find the container with id 96859d6a59b58c9df792a590deef50eb0ee923d03cb16fdc72abe3d18e466eaa Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.408068 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6snv6"] Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.416589 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4w5bf"] Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.442634 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpqvc"] Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.459200 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l9nqd" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.478700 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-4pbj2"] Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.478750 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vmwvg"] Jan 26 17:00:54 crc kubenswrapper[4856]: W0126 17:00:54.482935 4856 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfa11789e_7a2a_4dbf_85ca_c20a9d64a1f4.slice/crio-92cb1cfef16a0e75e3959ea9e1e938f0b67fd0c0c799ef25c376a0bf826c395e WatchSource:0}: Error finding container 92cb1cfef16a0e75e3959ea9e1e938f0b67fd0c0c799ef25c376a0bf826c395e: Status 404 returned error can't find the container with id 92cb1cfef16a0e75e3959ea9e1e938f0b67fd0c0c799ef25c376a0bf826c395e Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.497894 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.498219 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cb9fb12b-3eb8-4e94-a8cf-9eaf4703a850-cert\") pod \"ingress-canary-fbsj7\" (UID: \"cb9fb12b-3eb8-4e94-a8cf-9eaf4703a850\") " pod="openshift-ingress-canary/ingress-canary-fbsj7" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.498243 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cfa40861-cc08-4145-a185-6a3fb07eaabe-registry-certificates\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.498394 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cfa40861-cc08-4145-a185-6a3fb07eaabe-registry-tls\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.498421 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cfa40861-cc08-4145-a185-6a3fb07eaabe-ca-trust-extracted\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.498443 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cfa40861-cc08-4145-a185-6a3fb07eaabe-bound-sa-token\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.499237 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/113d2266-0e67-4e79-8a17-1a78cb9a13d5-metrics-tls\") pod \"dns-operator-744455d44c-z7cgq\" (UID: \"113d2266-0e67-4e79-8a17-1a78cb9a13d5\") " pod="openshift-dns-operator/dns-operator-744455d44c-z7cgq" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.499287 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cfa40861-cc08-4145-a185-6a3fb07eaabe-trusted-ca\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.499327 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tf448\" (UniqueName: \"kubernetes.io/projected/cfa40861-cc08-4145-a185-6a3fb07eaabe-kube-api-access-tf448\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.499868 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9tb9\" (UniqueName: \"kubernetes.io/projected/cb9fb12b-3eb8-4e94-a8cf-9eaf4703a850-kube-api-access-s9tb9\") pod \"ingress-canary-fbsj7\" (UID: \"cb9fb12b-3eb8-4e94-a8cf-9eaf4703a850\") " pod="openshift-ingress-canary/ingress-canary-fbsj7" Jan 26 17:00:54 crc kubenswrapper[4856]: E0126 17:00:54.499958 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:00:54.999940524 +0000 UTC m=+150.953194505 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.500009 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5q847\" (UniqueName: \"kubernetes.io/projected/113d2266-0e67-4e79-8a17-1a78cb9a13d5-kube-api-access-5q847\") pod \"dns-operator-744455d44c-z7cgq\" (UID: \"113d2266-0e67-4e79-8a17-1a78cb9a13d5\") " pod="openshift-dns-operator/dns-operator-744455d44c-z7cgq" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.500096 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cfa40861-cc08-4145-a185-6a3fb07eaabe-installation-pull-secrets\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.504411 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cfa40861-cc08-4145-a185-6a3fb07eaabe-ca-trust-extracted\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.509206 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cfa40861-cc08-4145-a185-6a3fb07eaabe-trusted-ca\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.510927 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cfa40861-cc08-4145-a185-6a3fb07eaabe-registry-certificates\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.513685 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/113d2266-0e67-4e79-8a17-1a78cb9a13d5-metrics-tls\") pod \"dns-operator-744455d44c-z7cgq\" (UID: \"113d2266-0e67-4e79-8a17-1a78cb9a13d5\") " pod="openshift-dns-operator/dns-operator-744455d44c-z7cgq" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.513797 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cfa40861-cc08-4145-a185-6a3fb07eaabe-registry-tls\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.516255 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cb9fb12b-3eb8-4e94-a8cf-9eaf4703a850-cert\") pod \"ingress-canary-fbsj7\" (UID: \"cb9fb12b-3eb8-4e94-a8cf-9eaf4703a850\") " pod="openshift-ingress-canary/ingress-canary-fbsj7" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.537322 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cfa40861-cc08-4145-a185-6a3fb07eaabe-installation-pull-secrets\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.540633 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-7l927"] Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.558107 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9tb9\" (UniqueName: \"kubernetes.io/projected/cb9fb12b-3eb8-4e94-a8cf-9eaf4703a850-kube-api-access-s9tb9\") pod \"ingress-canary-fbsj7\" (UID: \"cb9fb12b-3eb8-4e94-a8cf-9eaf4703a850\") " pod="openshift-ingress-canary/ingress-canary-fbsj7" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.566669 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cfa40861-cc08-4145-a185-6a3fb07eaabe-bound-sa-token\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.567043 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cl895"] Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.580077 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5q847\" (UniqueName: \"kubernetes.io/projected/113d2266-0e67-4e79-8a17-1a78cb9a13d5-kube-api-access-5q847\") pod \"dns-operator-744455d44c-z7cgq\" (UID: \"113d2266-0e67-4e79-8a17-1a78cb9a13d5\") " pod="openshift-dns-operator/dns-operator-744455d44c-z7cgq" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.597983 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7p5jt"] Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.601355 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mr7cp"] Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.601810 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tf448\" (UniqueName: \"kubernetes.io/projected/cfa40861-cc08-4145-a185-6a3fb07eaabe-kube-api-access-tf448\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.604384 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:54 crc kubenswrapper[4856]: E0126 17:00:54.604932 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:00:55.10491575 +0000 UTC m=+151.058169731 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.611452 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-cb8nk"] Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.707426 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:00:54 crc kubenswrapper[4856]: E0126 17:00:54.708834 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:00:55.207609818 +0000 UTC m=+151.160863799 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.728725 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:54 crc kubenswrapper[4856]: E0126 17:00:54.729429 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:00:55.229411991 +0000 UTC m=+151.182665982 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.732617 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-gz7kg"] Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.750321 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-7xb2b"] Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.759878 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-6cghs"] Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.827315 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-z7cgq" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.829706 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:00:54 crc kubenswrapper[4856]: E0126 17:00:54.829965 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:00:55.329941756 +0000 UTC m=+151.283195737 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.830373 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:54 crc kubenswrapper[4856]: E0126 17:00:54.831562 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:00:55.331538053 +0000 UTC m=+151.284792034 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.835105 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-fbsj7" Jan 26 17:00:54 crc kubenswrapper[4856]: I0126 17:00:54.932157 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:00:54 crc kubenswrapper[4856]: E0126 17:00:54.932751 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:00:55.432729387 +0000 UTC m=+151.385983378 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.034096 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:55 crc kubenswrapper[4856]: E0126 17:00:55.034687 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:00:55.534672333 +0000 UTC m=+151.487926314 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.134962 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:00:55 crc kubenswrapper[4856]: E0126 17:00:55.135397 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:00:55.635378093 +0000 UTC m=+151.588632074 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.183873 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-7xb2b" event={"ID":"77a97acb-2908-48fb-8bcd-0647f3e90160","Type":"ContainerStarted","Data":"71ba6cb40fc671d8e66509146fc662f55be8bb98bd417ad56358526161b87367"} Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.192440 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-dgcqn"] Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.210305 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-7l927" event={"ID":"94291fa4-24a5-499e-8143-89c8784d9284","Type":"ContainerStarted","Data":"82813129de322c75fa39cc94c94ea8625d6eed2d3ea3f14dd0db7911c9652bb4"} Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.227179 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8m4l6"] Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.233723 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-cjzsq"] Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.235743 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7p5jt" event={"ID":"bf1f11c8-17b8-49b7-b12d-92891f478222","Type":"ContainerStarted","Data":"847f817e4681c5f1a2ff6a9b03f573316d052751e866dfd065b55170125ec233"} Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.237931 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:55 crc kubenswrapper[4856]: E0126 17:00:55.238384 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:00:55.73837045 +0000 UTC m=+151.691624431 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.253597 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-27vjc" event={"ID":"5c244eff-aada-44f3-b250-96878a3400c4","Type":"ContainerStarted","Data":"27da60ab04d9c096ac5db9ea266e1bcf3e305705808f89e734cdc0d040595272"} Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.259592 4856 generic.go:334] "Generic (PLEG): container finished" podID="2ba3cf6a-a6be-4108-a155-c8bb530aa037" containerID="d89a3287b371ef0cb23bab1eb475eccd1999b48798e71ba74c19f952107aef34" exitCode=0 Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.259682 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bjl7" event={"ID":"2ba3cf6a-a6be-4108-a155-c8bb530aa037","Type":"ContainerDied","Data":"d89a3287b371ef0cb23bab1eb475eccd1999b48798e71ba74c19f952107aef34"} Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.273506 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6cghs" event={"ID":"81c2f96b-55e0-483b-b72c-df7e156e9218","Type":"ContainerStarted","Data":"dac4ba81d1cdc88dc980c2ceb845187c7a1ed41d1d2eec1cf749f33ac5b8b442"} Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.278198 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cl895" event={"ID":"42c0e428-821f-45a1-85a7-54ebdb81ef1c","Type":"ContainerStarted","Data":"cec55d4172500b35f1678464fe1ce0649bb48ca145dea45b1d9baeaa952e2041"} Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.285081 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mr7cp" event={"ID":"f9b8f57e-00b9-4355-ace2-0319d320d208","Type":"ContainerStarted","Data":"5323c8c9fdb9e4225fb31744181dc0b0bd41776e30729fe779fc73fb5659e9a6"} Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.288690 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-h9b2g" event={"ID":"85f05bd5-ff83-4d29-9531-ab3499088095","Type":"ContainerStarted","Data":"1747ce02dc365e25bdb1e14cb852860f0ec5220a32a63437c450ff9da4361ed5"} Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.293601 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-gz7kg" event={"ID":"033cb12f-278f-431a-8104-519db9a3152f","Type":"ContainerStarted","Data":"361fe6998603fd54ac5cfe0959e0cbc545d3c5e143e033477253dcf3d57d7a23"} Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.297840 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-88lkr" event={"ID":"359660cd-b412-4640-bedf-993e976e7b3c","Type":"ContainerStarted","Data":"fe76caa8fee51b66026c972b86975afe11768c8e99f4e26a8014d26d3187b8d4"} Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.300630 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-c9qdp" event={"ID":"437b5573-a342-4383-ba60-be0e3ccba839","Type":"ContainerStarted","Data":"5703d26d139036e12af52a438fa520c387d8a24070f61cff00688fb6c5224867"} Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.304710 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-jdjcq" event={"ID":"a7e5d16a-d45d-4b40-93bf-bfaa6be2d1c0","Type":"ContainerStarted","Data":"ef1cb2622920f923afb73fc12eda34877038d999d31cb4114611a29f8ca1fbd5"} Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.314657 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vmwvg" event={"ID":"ddc2e6b7-5582-4579-bf2c-ed165b74c91a","Type":"ContainerStarted","Data":"9c8b75ffa9d72b626a3cbee0eb9647978bd9abaf50267db3c9990debab4058e3"} Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.317693 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" event={"ID":"a6d331bd-2db3-4319-9f5c-db56d408d9e3","Type":"ContainerStarted","Data":"bae2856860ccbfbac0017e7da1d4f30234d58daef19aaf9d48d7644f9aac2b38"} Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.319350 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6snv6" event={"ID":"fa11789e-7a2a-4dbf-85ca-c20a9d64a1f4","Type":"ContainerStarted","Data":"92cb1cfef16a0e75e3959ea9e1e938f0b67fd0c0c799ef25c376a0bf826c395e"} Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.322318 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-962cr" event={"ID":"54a246a2-f674-4735-b295-b56699ece95b","Type":"ContainerStarted","Data":"2953035f4fdd7266bb0ef7eba50a3cc88c7ef83c9d339275478ed0a1da8e092e"} Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.323273 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" event={"ID":"69008ed1-f3e5-400d-852f-adbcd94199f6","Type":"ContainerStarted","Data":"d2e5352f5a4f0bdf4461c4b926a9353c0b4a673c6263c30adba1a3d7a2d6a8ad"} Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.324052 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpqvc" event={"ID":"5fe6baed-ab97-4d8a-8be2-6f00f9698136","Type":"ContainerStarted","Data":"e3a4f0c156036789efac8b4cdbd3ace5dcdaf8c187d261687c3b9c87a15d74df"} Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.325578 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-lndnt" event={"ID":"1afc0f4c-e02d-4a70-aaba-e761e8c04eee","Type":"ContainerStarted","Data":"96859d6a59b58c9df792a590deef50eb0ee923d03cb16fdc72abe3d18e466eaa"} Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.326093 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-lndnt" Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.328367 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-q7j7b" event={"ID":"149e3000-35d7-47bd-83f0-00ab5e0736c2","Type":"ContainerStarted","Data":"b8e5f5596ec8c94d661b87d0c08b723c1088f87879918a4b4f2e2a14cb1358fb"} Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.340427 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:00:55 crc kubenswrapper[4856]: E0126 17:00:55.340780 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:00:55.840747919 +0000 UTC m=+151.794001900 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.340427 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-6qgnn" event={"ID":"b28404ed-2e71-4b3f-9140-35ee89dbc8f2","Type":"ContainerStarted","Data":"cf3a5cf0d543759d7e8e6e68a6bd4c1b71efee63ebda8e6116e343b588f4f9f9"} Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.341580 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:55 crc kubenswrapper[4856]: E0126 17:00:55.342178 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:00:55.8421554 +0000 UTC m=+151.795409441 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.343083 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-4pbj2" event={"ID":"831dc87e-8e14-43d3-a36e-dc7679041ae5","Type":"ContainerStarted","Data":"65096e9aeb5bb4bd72bc2e5027f15d210d5133a05fe55fb568baf6443a93e32c"} Jan 26 17:00:55 crc kubenswrapper[4856]: W0126 17:00:55.437917 4856 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac10f013_cd1f_47e0_8f1c_5ff4e6e75784.slice/crio-b81ebf6bd6b806c8fa7a6e0ba1caa582537dbc1e86b1c60f8089719a1ee5c590 WatchSource:0}: Error finding container b81ebf6bd6b806c8fa7a6e0ba1caa582537dbc1e86b1c60f8089719a1ee5c590: Status 404 returned error can't find the container with id b81ebf6bd6b806c8fa7a6e0ba1caa582537dbc1e86b1c60f8089719a1ee5c590 Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.442601 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:00:55 crc kubenswrapper[4856]: E0126 17:00:55.442739 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:00:55.942710726 +0000 UTC m=+151.895964707 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.443136 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:55 crc kubenswrapper[4856]: E0126 17:00:55.444808 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:00:55.944794147 +0000 UTC m=+151.898048128 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.461580 4856 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-lndnt container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Jan 26 17:00:55 crc kubenswrapper[4856]: W0126 17:00:55.461749 4856 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod550752e4_a1d9_46f4_9118_9e9919b2fe6b.slice/crio-582a04946abdbc0330959e567196855ede2ec9ec2977ecc42023606fb6a2ddfe WatchSource:0}: Error finding container 582a04946abdbc0330959e567196855ede2ec9ec2977ecc42023606fb6a2ddfe: Status 404 returned error can't find the container with id 582a04946abdbc0330959e567196855ede2ec9ec2977ecc42023606fb6a2ddfe Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.462905 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-lndnt" podUID="1afc0f4c-e02d-4a70-aaba-e761e8c04eee" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.545647 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:00:55 crc kubenswrapper[4856]: E0126 17:00:55.545939 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:00:56.045902108 +0000 UTC m=+151.999156089 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.546312 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:55 crc kubenswrapper[4856]: E0126 17:00:55.546689 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:00:56.046677651 +0000 UTC m=+151.999931632 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.907390 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:00:55 crc kubenswrapper[4856]: E0126 17:00:55.907846 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:00:56.407824291 +0000 UTC m=+152.361078272 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.960487 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l9nqd"] Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.960577 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nn46h"] Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.971314 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-ddghz"] Jan 26 17:00:55 crc kubenswrapper[4856]: I0126 17:00:55.978566 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k662z"] Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.000446 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rrhjv"] Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.008653 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:56 crc kubenswrapper[4856]: E0126 17:00:56.009065 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:00:56.509046616 +0000 UTC m=+152.462300597 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:56 crc kubenswrapper[4856]: W0126 17:00:56.015898 4856 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbeb6f283_75cb_4184_b985_4e6c095feca1.slice/crio-e86e9d9dead0ba26d06a2376812c403d4aa6815fc1e8e161b932b21c3de11e00 WatchSource:0}: Error finding container e86e9d9dead0ba26d06a2376812c403d4aa6815fc1e8e161b932b21c3de11e00: Status 404 returned error can't find the container with id e86e9d9dead0ba26d06a2376812c403d4aa6815fc1e8e161b932b21c3de11e00 Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.039285 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wvttb"] Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.155971 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:00:56 crc kubenswrapper[4856]: E0126 17:00:56.156463 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:00:56.656442753 +0000 UTC m=+152.609696734 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.194086 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490780-8q6q4"] Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.197569 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-2sfhr"] Jan 26 17:00:56 crc kubenswrapper[4856]: W0126 17:00:56.208690 4856 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda4d83db5_776f_4e95_a6fa_b194344f9819.slice/crio-0a33ee036611e9639534d8e4446ef3e1eb207d3627ffb8748d14c0cdf8b384ce WatchSource:0}: Error finding container 0a33ee036611e9639534d8e4446ef3e1eb207d3627ffb8748d14c0cdf8b384ce: Status 404 returned error can't find the container with id 0a33ee036611e9639534d8e4446ef3e1eb207d3627ffb8748d14c0cdf8b384ce Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.252042 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-qdmxz"] Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.261845 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:56 crc kubenswrapper[4856]: E0126 17:00:56.262175 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:00:56.76216125 +0000 UTC m=+152.715415231 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.265997 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-vfm8t"] Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.292731 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-6qgnn" podStartSLOduration=122.292704791 podStartE2EDuration="2m2.292704791s" podCreationTimestamp="2026-01-26 16:58:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:00:56.290247958 +0000 UTC m=+152.243501959" watchObservedRunningTime="2026-01-26 17:00:56.292704791 +0000 UTC m=+152.245958772" Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.322757 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-q7j7b" podStartSLOduration=121.322727286 podStartE2EDuration="2m1.322727286s" podCreationTimestamp="2026-01-26 16:58:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:00:56.316210144 +0000 UTC m=+152.269464135" watchObservedRunningTime="2026-01-26 17:00:56.322727286 +0000 UTC m=+152.275981277" Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.324742 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-zzxln"] Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.338836 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-fbsj7"] Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.341924 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-88lkr" podStartSLOduration=123.341904002 podStartE2EDuration="2m3.341904002s" podCreationTimestamp="2026-01-26 16:58:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:00:56.340283504 +0000 UTC m=+152.293537495" watchObservedRunningTime="2026-01-26 17:00:56.341904002 +0000 UTC m=+152.295157983" Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.360164 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-58fcz"] Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.362718 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-z7cgq"] Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.368330 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:00:56 crc kubenswrapper[4856]: E0126 17:00:56.368923 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:00:56.868903718 +0000 UTC m=+152.822157699 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.371276 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mr7cp" event={"ID":"f9b8f57e-00b9-4355-ace2-0319d320d208","Type":"ContainerStarted","Data":"e4bcd086b81f4285b2ca3164dcb88913e0abc6f6f9c8ae1d8f771b21afd48202"} Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.374562 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4w5bf" event={"ID":"0c1af7db-aa80-4cb0-a9cb-5afdf677f28c","Type":"ContainerStarted","Data":"ad19be306a94469ab15b9268231447bdbbb8283c3860820099b42bbc7e87b980"} Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.376394 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-jdjcq" event={"ID":"a7e5d16a-d45d-4b40-93bf-bfaa6be2d1c0","Type":"ContainerStarted","Data":"769ebacaf7d061c49dd62c73c9ee5eb8d4bcc193b6c7cde28085a6dd9765f5e9"} Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.382913 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-c9qdp" event={"ID":"437b5573-a342-4383-ba60-be0e3ccba839","Type":"ContainerStarted","Data":"ec0e040a6eff7a45d3bb88143a951575d5e006f1f8e3c81af1602d243e444da4"} Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.386814 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-jdjcq" podStartSLOduration=123.386794866 podStartE2EDuration="2m3.386794866s" podCreationTimestamp="2026-01-26 16:58:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:00:56.385897949 +0000 UTC m=+152.339151950" watchObservedRunningTime="2026-01-26 17:00:56.386794866 +0000 UTC m=+152.340048847" Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.388199 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-lndnt" event={"ID":"1afc0f4c-e02d-4a70-aaba-e761e8c04eee","Type":"ContainerStarted","Data":"e9e54e2a4a2266ca4148b11cb38df08f87c2f2ccd87dc3343d147862786c16e2"} Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.392737 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-dgcqn" event={"ID":"ac10f013-cd1f-47e0-8f1c-5ff4e6e75784","Type":"ContainerStarted","Data":"b81ebf6bd6b806c8fa7a6e0ba1caa582537dbc1e86b1c60f8089719a1ee5c590"} Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.396914 4856 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-lndnt container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.396977 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-lndnt" podUID="1afc0f4c-e02d-4a70-aaba-e761e8c04eee" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.409861 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-lndnt" podStartSLOduration=122.409844735 podStartE2EDuration="2m2.409844735s" podCreationTimestamp="2026-01-26 16:58:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:00:56.409066362 +0000 UTC m=+152.362320373" watchObservedRunningTime="2026-01-26 17:00:56.409844735 +0000 UTC m=+152.363098716" Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.410338 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-962cr" event={"ID":"54a246a2-f674-4735-b295-b56699ece95b","Type":"ContainerStarted","Data":"17f583a2b977ca8d5b98468ca5a47ddf673a2a5ac93b0240f5ad0498f9ab0f38"} Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.411626 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2sfhr" event={"ID":"a4d83db5-776f-4e95-a6fa-b194344f9819","Type":"ContainerStarted","Data":"0a33ee036611e9639534d8e4446ef3e1eb207d3627ffb8748d14c0cdf8b384ce"} Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.412576 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wvttb" event={"ID":"2d37efbf-d18f-486b-9b43-bc4d181af4ca","Type":"ContainerStarted","Data":"fff8ee4c0db342e8c666d6319a47d7101521fb44435e8030d5a5dc565b0b6c44"} Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.413491 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8m4l6" event={"ID":"004316da-16cd-49ab-b14d-282c28da6fad","Type":"ContainerStarted","Data":"b8122b481c09344feb679c573e930e25f1622bd18aee7661079598c6d3a9a8ac"} Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.414464 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qdmxz" event={"ID":"37a77f41-5dbf-4842-9e77-83dc22b50f4a","Type":"ContainerStarted","Data":"af1abc94505937923c6569d8078b91b7d858b20669a62b1e4e2df83f5cf3159c"} Jan 26 17:00:56 crc kubenswrapper[4856]: W0126 17:00:56.417963 4856 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb9fb12b_3eb8_4e94_a8cf_9eaf4703a850.slice/crio-d0b72852c2a3bf2085df9dba8e4da15ce763cef156a346cec23a9847b6fb31d9 WatchSource:0}: Error finding container d0b72852c2a3bf2085df9dba8e4da15ce763cef156a346cec23a9847b6fb31d9: Status 404 returned error can't find the container with id d0b72852c2a3bf2085df9dba8e4da15ce763cef156a346cec23a9847b6fb31d9 Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.420368 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nn46h" event={"ID":"abbeffe1-cfd5-4476-9a8e-2ab5b4869444","Type":"ContainerStarted","Data":"4db554ffddfe2417dd5add45dccd6872bfa1f92a542c1e144f3c8a956cc95996"} Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.425804 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-c9qdp" podStartSLOduration=5.425788996 podStartE2EDuration="5.425788996s" podCreationTimestamp="2026-01-26 17:00:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:00:56.423268071 +0000 UTC m=+152.376522062" watchObservedRunningTime="2026-01-26 17:00:56.425788996 +0000 UTC m=+152.379042977" Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.431017 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-cjzsq" event={"ID":"550752e4-a1d9-46f4-9118-9e9919b2fe6b","Type":"ContainerStarted","Data":"582a04946abdbc0330959e567196855ede2ec9ec2977ecc42023606fb6a2ddfe"} Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.439587 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-ddghz" event={"ID":"beb6f283-75cb-4184-b985-4e6c095feca1","Type":"ContainerStarted","Data":"e86e9d9dead0ba26d06a2376812c403d4aa6815fc1e8e161b932b21c3de11e00"} Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.447495 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rrhjv" event={"ID":"a1546392-4a69-4b12-8d7e-97450b73b7ca","Type":"ContainerStarted","Data":"6d06091dc3fc59968137d394ab73397c735381fe527eabe19085692a3a73391f"} Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.450904 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-q7j7b" event={"ID":"149e3000-35d7-47bd-83f0-00ab5e0736c2","Type":"ContainerStarted","Data":"36822ae5f9ab805c23d325fa959ca3921a74a99a02a6f3f4c0fa39de2050fd6e"} Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.452109 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-8q6q4" event={"ID":"7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7","Type":"ContainerStarted","Data":"d8fe561f33f411cab54065acf50663e1fea5f5209ab612f88976297cc920acef"} Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.461678 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l9nqd" event={"ID":"05d74105-0ecd-41ac-9001-8b21b0fd6ba4","Type":"ContainerStarted","Data":"109ffd7506918ece16ce7aa025aeab703d068dcf1cd5382a1665ed919d9947c7"} Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.466471 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k662z" event={"ID":"17a72e73-4d54-4a29-a85a-ecb1aff30d10","Type":"ContainerStarted","Data":"3cda7b649b2ddbe51156afa828e1b3ce94f39d3ac82bca94653c80e65d76c66f"} Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.470252 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:56 crc kubenswrapper[4856]: E0126 17:00:56.473489 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:00:56.973473452 +0000 UTC m=+152.926727433 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.572648 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:00:56 crc kubenswrapper[4856]: E0126 17:00:56.573723 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:00:57.073687487 +0000 UTC m=+153.026941478 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.675104 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:56 crc kubenswrapper[4856]: E0126 17:00:56.675486 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:00:57.175471289 +0000 UTC m=+153.128725270 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.775954 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:00:56 crc kubenswrapper[4856]: E0126 17:00:56.776130 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:00:57.276106556 +0000 UTC m=+153.229360537 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.776283 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:56 crc kubenswrapper[4856]: E0126 17:00:56.776615 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:00:57.276606001 +0000 UTC m=+153.229859982 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.877812 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:00:56 crc kubenswrapper[4856]: E0126 17:00:56.878218 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:00:57.378179696 +0000 UTC m=+153.331433677 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.950390 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 17:00:56 crc kubenswrapper[4856]: I0126 17:00:56.980045 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:56 crc kubenswrapper[4856]: E0126 17:00:56.980465 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:00:57.480451962 +0000 UTC m=+153.433705943 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:57 crc kubenswrapper[4856]: I0126 17:00:57.082702 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:00:57 crc kubenswrapper[4856]: E0126 17:00:57.082928 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:00:57.582892173 +0000 UTC m=+153.536146154 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:57 crc kubenswrapper[4856]: I0126 17:00:57.084931 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:57 crc kubenswrapper[4856]: E0126 17:00:57.085430 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:00:57.585419358 +0000 UTC m=+153.538673419 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:57 crc kubenswrapper[4856]: I0126 17:00:57.186500 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:00:57 crc kubenswrapper[4856]: E0126 17:00:57.186832 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:00:57.686789737 +0000 UTC m=+153.640043728 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:57 crc kubenswrapper[4856]: I0126 17:00:57.187071 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:57 crc kubenswrapper[4856]: E0126 17:00:57.187713 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:00:57.687687513 +0000 UTC m=+153.640941494 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:57 crc kubenswrapper[4856]: I0126 17:00:57.290615 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:00:57 crc kubenswrapper[4856]: E0126 17:00:57.290885 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:00:57.790824315 +0000 UTC m=+153.744078296 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:57 crc kubenswrapper[4856]: I0126 17:00:57.291198 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:57 crc kubenswrapper[4856]: E0126 17:00:57.291580 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:00:57.791564447 +0000 UTC m=+153.744818428 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:57 crc kubenswrapper[4856]: I0126 17:00:57.392755 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:00:57 crc kubenswrapper[4856]: E0126 17:00:57.393102 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:00:57.89306607 +0000 UTC m=+153.846320061 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:57 crc kubenswrapper[4856]: I0126 17:00:57.393358 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:57 crc kubenswrapper[4856]: E0126 17:00:57.393773 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:00:57.89374421 +0000 UTC m=+153.846998191 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:57 crc kubenswrapper[4856]: I0126 17:00:57.472184 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-fbsj7" event={"ID":"cb9fb12b-3eb8-4e94-a8cf-9eaf4703a850","Type":"ContainerStarted","Data":"d0b72852c2a3bf2085df9dba8e4da15ce763cef156a346cec23a9847b6fb31d9"} Jan 26 17:00:57 crc kubenswrapper[4856]: I0126 17:00:57.473497 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-27vjc" event={"ID":"5c244eff-aada-44f3-b250-96878a3400c4","Type":"ContainerStarted","Data":"9ba6ea817a3c31b0e39f2111b57da282eb2b77c13e0ffcfcc27bac01ba4cc371"} Jan 26 17:00:57 crc kubenswrapper[4856]: I0126 17:00:57.474390 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-z7cgq" event={"ID":"113d2266-0e67-4e79-8a17-1a78cb9a13d5","Type":"ContainerStarted","Data":"edbf7d4a874910b774b68f7bbe85478e92a569aff5448e9bbe0b693dd010ecbd"} Jan 26 17:00:57 crc kubenswrapper[4856]: I0126 17:00:57.475678 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-58fcz" event={"ID":"73de6ef2-e139-4185-9f56-9db885734ffe","Type":"ContainerStarted","Data":"7b62cf3316521561ce0b38913d1a2fee49f3f67652607f4d7ede0d1bd6bdef3f"} Jan 26 17:00:57 crc kubenswrapper[4856]: I0126 17:00:57.476456 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-vfm8t" event={"ID":"c8657575-cd22-4ebc-ae9d-4174366985d3","Type":"ContainerStarted","Data":"f5c708af576ac7ab49284d96d259123abba4d963b1a82c6a53c2a3bf5852203e"} Jan 26 17:00:57 crc kubenswrapper[4856]: I0126 17:00:57.477439 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-zzxln" event={"ID":"129a0b30-7132-4e3c-ab84-208cae7cb2f2","Type":"ContainerStarted","Data":"610adfd4db40083b842818daf5fbf443f006512d8a12c35a52868753c3132f1e"} Jan 26 17:00:57 crc kubenswrapper[4856]: I0126 17:00:57.478839 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" event={"ID":"a6d331bd-2db3-4319-9f5c-db56d408d9e3","Type":"ContainerStarted","Data":"c693764cd154e71eb16d8ca854be70839609355fd70555fd1faebdfa8a4e3e40"} Jan 26 17:00:57 crc kubenswrapper[4856]: I0126 17:00:57.480262 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6snv6" event={"ID":"fa11789e-7a2a-4dbf-85ca-c20a9d64a1f4","Type":"ContainerStarted","Data":"70c82482161b3663e98d8206f317df90be3d8ee251bd8ac00e9b981b55d6156b"} Jan 26 17:00:57 crc kubenswrapper[4856]: I0126 17:00:57.481054 4856 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-lndnt container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Jan 26 17:00:57 crc kubenswrapper[4856]: I0126 17:00:57.481095 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-lndnt" podUID="1afc0f4c-e02d-4a70-aaba-e761e8c04eee" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Jan 26 17:00:57 crc kubenswrapper[4856]: I0126 17:00:57.494594 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:00:57 crc kubenswrapper[4856]: E0126 17:00:57.494718 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:00:57.994692797 +0000 UTC m=+153.947946788 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:57 crc kubenswrapper[4856]: I0126 17:00:57.494887 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:57 crc kubenswrapper[4856]: E0126 17:00:57.495330 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:00:57.995310955 +0000 UTC m=+153.948564936 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:57 crc kubenswrapper[4856]: I0126 17:00:57.595680 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:00:57 crc kubenswrapper[4856]: E0126 17:00:57.595959 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:00:58.095931522 +0000 UTC m=+154.049185503 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:57 crc kubenswrapper[4856]: I0126 17:00:57.596423 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:57 crc kubenswrapper[4856]: E0126 17:00:57.603210 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:00:58.103142345 +0000 UTC m=+154.056396326 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:57 crc kubenswrapper[4856]: I0126 17:00:57.697512 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:00:57 crc kubenswrapper[4856]: E0126 17:00:57.698088 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:00:58.198068674 +0000 UTC m=+154.151322655 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:57 crc kubenswrapper[4856]: I0126 17:00:57.799510 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:57 crc kubenswrapper[4856]: E0126 17:00:57.800171 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:00:58.300140174 +0000 UTC m=+154.253394335 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:57 crc kubenswrapper[4856]: I0126 17:00:57.900856 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:00:57 crc kubenswrapper[4856]: E0126 17:00:57.901229 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:00:58.401196835 +0000 UTC m=+154.354450826 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:58 crc kubenswrapper[4856]: I0126 17:00:58.002380 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:58 crc kubenswrapper[4856]: E0126 17:00:58.002885 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:00:58.502864433 +0000 UTC m=+154.456118414 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:58 crc kubenswrapper[4856]: I0126 17:00:58.104946 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:00:58 crc kubenswrapper[4856]: E0126 17:00:58.105288 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:00:58.605259702 +0000 UTC m=+154.558513713 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:58 crc kubenswrapper[4856]: I0126 17:00:58.206946 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:58 crc kubenswrapper[4856]: E0126 17:00:58.207418 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:00:58.707397834 +0000 UTC m=+154.660651815 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:58 crc kubenswrapper[4856]: I0126 17:00:58.308184 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:00:58 crc kubenswrapper[4856]: E0126 17:00:58.308436 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:00:58.808404373 +0000 UTC m=+154.761658354 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:58 crc kubenswrapper[4856]: I0126 17:00:58.308619 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:58 crc kubenswrapper[4856]: E0126 17:00:58.309025 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:00:58.809008621 +0000 UTC m=+154.762262602 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:58 crc kubenswrapper[4856]: I0126 17:00:58.409368 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:00:58 crc kubenswrapper[4856]: E0126 17:00:58.409861 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:00:58.909824204 +0000 UTC m=+154.863078195 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:58 crc kubenswrapper[4856]: I0126 17:00:58.410057 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:58 crc kubenswrapper[4856]: E0126 17:00:58.410500 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:00:58.910483683 +0000 UTC m=+154.863737724 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:58 crc kubenswrapper[4856]: I0126 17:00:58.519573 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:00:58 crc kubenswrapper[4856]: E0126 17:00:58.519975 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:00:59.019957161 +0000 UTC m=+154.973211142 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:58 crc kubenswrapper[4856]: I0126 17:00:58.638904 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:58 crc kubenswrapper[4856]: E0126 17:00:58.639810 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:00:59.139785495 +0000 UTC m=+155.093039476 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:58 crc kubenswrapper[4856]: I0126 17:00:58.678452 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6cghs" event={"ID":"81c2f96b-55e0-483b-b72c-df7e156e9218","Type":"ContainerStarted","Data":"23daacc6ab399456fe30d31494993067d0f18366b56b87411ba96377ebdd2807"} Jan 26 17:00:58 crc kubenswrapper[4856]: I0126 17:00:58.680354 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" event={"ID":"69008ed1-f3e5-400d-852f-adbcd94199f6","Type":"ContainerStarted","Data":"749ef964d6b168f431c27d0286b92e40d64a8b4fb99f430b33432827ee871fc9"} Jan 26 17:00:58 crc kubenswrapper[4856]: I0126 17:00:58.681132 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:00:58 crc kubenswrapper[4856]: I0126 17:00:58.692004 4856 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-cb8nk container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" start-of-body= Jan 26 17:00:58 crc kubenswrapper[4856]: I0126 17:00:58.692075 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" podUID="69008ed1-f3e5-400d-852f-adbcd94199f6" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" Jan 26 17:00:58 crc kubenswrapper[4856]: I0126 17:00:58.694445 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bjl7" event={"ID":"2ba3cf6a-a6be-4108-a155-c8bb530aa037","Type":"ContainerStarted","Data":"7696af9bf5eb7a27c45bc9a500fea17921f66464546a3b193d1abfd56ccd50c4"} Jan 26 17:00:58 crc kubenswrapper[4856]: I0126 17:00:58.742311 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:00:58 crc kubenswrapper[4856]: E0126 17:00:58.742889 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:00:59.242862255 +0000 UTC m=+155.196116246 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:58 crc kubenswrapper[4856]: I0126 17:00:58.755455 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cl895" event={"ID":"42c0e428-821f-45a1-85a7-54ebdb81ef1c","Type":"ContainerStarted","Data":"95b79a7170c279e8c4d88caa46ddb2c7a788acc3cffb03a4d607b68a5fd42fc4"} Jan 26 17:00:58 crc kubenswrapper[4856]: I0126 17:00:58.775853 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4w5bf" event={"ID":"0c1af7db-aa80-4cb0-a9cb-5afdf677f28c","Type":"ContainerStarted","Data":"3609c7e69071336f3f98579608ca0b0b398d8bcf3cebb1a6686cd170813984bf"} Jan 26 17:00:58 crc kubenswrapper[4856]: I0126 17:00:58.844479 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:58 crc kubenswrapper[4856]: E0126 17:00:58.873363 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:00:59.373326212 +0000 UTC m=+155.326580193 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:58 crc kubenswrapper[4856]: I0126 17:00:58.926117 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" podStartSLOduration=125.926086918 podStartE2EDuration="2m5.926086918s" podCreationTimestamp="2026-01-26 16:58:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:00:58.714000214 +0000 UTC m=+154.667254205" watchObservedRunningTime="2026-01-26 17:00:58.926086918 +0000 UTC m=+154.879340899" Jan 26 17:00:58 crc kubenswrapper[4856]: I0126 17:00:58.927291 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-cl895" podStartSLOduration=123.927284453 podStartE2EDuration="2m3.927284453s" podCreationTimestamp="2026-01-26 16:58:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:00:58.925266964 +0000 UTC m=+154.878520955" watchObservedRunningTime="2026-01-26 17:00:58.927284453 +0000 UTC m=+154.880538434" Jan 26 17:00:58 crc kubenswrapper[4856]: I0126 17:00:58.946650 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:00:58 crc kubenswrapper[4856]: E0126 17:00:58.947024 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:00:59.446989555 +0000 UTC m=+155.400243536 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:58 crc kubenswrapper[4856]: I0126 17:00:58.959519 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-7xb2b" event={"ID":"77a97acb-2908-48fb-8bcd-0647f3e90160","Type":"ContainerStarted","Data":"0a57b717a646c106b6a7d1b2a3fe85a7d08effd05da3065545df906392710c90"} Jan 26 17:00:58 crc kubenswrapper[4856]: I0126 17:00:58.961254 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7p5jt" event={"ID":"bf1f11c8-17b8-49b7-b12d-92891f478222","Type":"ContainerStarted","Data":"08d92ff7b7c49b1fb5dee4e1807177a3ef5855778282ac9d1949c687b64826ac"} Jan 26 17:00:58 crc kubenswrapper[4856]: I0126 17:00:58.965919 4856 generic.go:334] "Generic (PLEG): container finished" podID="a6d331bd-2db3-4319-9f5c-db56d408d9e3" containerID="c693764cd154e71eb16d8ca854be70839609355fd70555fd1faebdfa8a4e3e40" exitCode=0 Jan 26 17:00:58 crc kubenswrapper[4856]: I0126 17:00:58.965988 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" event={"ID":"a6d331bd-2db3-4319-9f5c-db56d408d9e3","Type":"ContainerDied","Data":"c693764cd154e71eb16d8ca854be70839609355fd70555fd1faebdfa8a4e3e40"} Jan 26 17:00:59 crc kubenswrapper[4856]: I0126 17:00:59.044641 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6snv6" podStartSLOduration=124.044622524 podStartE2EDuration="2m4.044622524s" podCreationTimestamp="2026-01-26 16:58:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:00:59.042852732 +0000 UTC m=+154.996106713" watchObservedRunningTime="2026-01-26 17:00:59.044622524 +0000 UTC m=+154.997876505" Jan 26 17:00:59 crc kubenswrapper[4856]: I0126 17:00:59.050933 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:59 crc kubenswrapper[4856]: E0126 17:00:59.054835 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:00:59.554808744 +0000 UTC m=+155.508062885 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:59 crc kubenswrapper[4856]: I0126 17:00:59.181171 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:00:59 crc kubenswrapper[4856]: E0126 17:00:59.182118 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:00:59.682097377 +0000 UTC m=+155.635351368 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:59 crc kubenswrapper[4856]: I0126 17:00:59.339748 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:59 crc kubenswrapper[4856]: E0126 17:00:59.340303 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:00:59.840278902 +0000 UTC m=+155.793532883 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:59 crc kubenswrapper[4856]: I0126 17:00:59.351649 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mr7cp" podStartSLOduration=124.351629636 podStartE2EDuration="2m4.351629636s" podCreationTimestamp="2026-01-26 16:58:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:00:59.351047949 +0000 UTC m=+155.304301950" watchObservedRunningTime="2026-01-26 17:00:59.351629636 +0000 UTC m=+155.304883637" Jan 26 17:00:59 crc kubenswrapper[4856]: I0126 17:00:59.441476 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:00:59 crc kubenswrapper[4856]: E0126 17:00:59.441999 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:00:59.94196887 +0000 UTC m=+155.895222861 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:59 crc kubenswrapper[4856]: I0126 17:00:59.559276 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:59 crc kubenswrapper[4856]: E0126 17:00:59.559898 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:00.059885268 +0000 UTC m=+156.013139249 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:59 crc kubenswrapper[4856]: I0126 17:00:59.733148 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:00:59 crc kubenswrapper[4856]: E0126 17:00:59.733940 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:00.233891039 +0000 UTC m=+156.187145040 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:00:59 crc kubenswrapper[4856]: I0126 17:00:59.837545 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:00:59 crc kubenswrapper[4856]: E0126 17:00:59.838174 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:00.338149904 +0000 UTC m=+156.291403885 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:00 crc kubenswrapper[4856]: I0126 17:01:00.052151 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:00 crc kubenswrapper[4856]: E0126 17:01:00.052345 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:00.55232351 +0000 UTC m=+156.505577491 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:00 crc kubenswrapper[4856]: I0126 17:01:00.052894 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:00 crc kubenswrapper[4856]: E0126 17:01:00.053256 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:00.553244627 +0000 UTC m=+156.506498608 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:00 crc kubenswrapper[4856]: I0126 17:01:00.379626 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:00 crc kubenswrapper[4856]: E0126 17:01:00.380100 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:00.880079035 +0000 UTC m=+156.833333016 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:00 crc kubenswrapper[4856]: I0126 17:01:00.427171 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-dgcqn" event={"ID":"ac10f013-cd1f-47e0-8f1c-5ff4e6e75784","Type":"ContainerStarted","Data":"babaad9608eea9b6b1dc555762da5c4716f4dc2f9aefbe5d18813e355824a597"} Jan 26 17:01:00 crc kubenswrapper[4856]: I0126 17:01:00.434466 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vmwvg" event={"ID":"ddc2e6b7-5582-4579-bf2c-ed165b74c91a","Type":"ContainerStarted","Data":"8c85eabf2567394a69674e3e201bd137e5b390af6c44f28f3472f530c44ae4b3"} Jan 26 17:01:00 crc kubenswrapper[4856]: I0126 17:01:00.444058 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-fbsj7" event={"ID":"cb9fb12b-3eb8-4e94-a8cf-9eaf4703a850","Type":"ContainerStarted","Data":"e9c57c3116c91bf8686961596571d2f04ec3432446e43013955f1874993811ea"} Jan 26 17:01:00 crc kubenswrapper[4856]: I0126 17:01:00.453311 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k662z" event={"ID":"17a72e73-4d54-4a29-a85a-ecb1aff30d10","Type":"ContainerStarted","Data":"f3f66e19dd32d40a6256207f49ffc3a9d666ce278b05bb4ef515b5be01f0d91a"} Jan 26 17:01:00 crc kubenswrapper[4856]: I0126 17:01:00.454614 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k662z" Jan 26 17:01:00 crc kubenswrapper[4856]: I0126 17:01:00.465231 4856 generic.go:334] "Generic (PLEG): container finished" podID="81c2f96b-55e0-483b-b72c-df7e156e9218" containerID="23daacc6ab399456fe30d31494993067d0f18366b56b87411ba96377ebdd2807" exitCode=0 Jan 26 17:01:00 crc kubenswrapper[4856]: I0126 17:01:00.465569 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6cghs" event={"ID":"81c2f96b-55e0-483b-b72c-df7e156e9218","Type":"ContainerDied","Data":"23daacc6ab399456fe30d31494993067d0f18366b56b87411ba96377ebdd2807"} Jan 26 17:01:00 crc kubenswrapper[4856]: I0126 17:01:00.470039 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-4pbj2" event={"ID":"831dc87e-8e14-43d3-a36e-dc7679041ae5","Type":"ContainerStarted","Data":"21bcecc04eac77160f1042631d12a9435ed2f19641a1d340f22c788470d2db4c"} Jan 26 17:01:00 crc kubenswrapper[4856]: I0126 17:01:00.471017 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-4pbj2" Jan 26 17:01:00 crc kubenswrapper[4856]: I0126 17:01:00.472833 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-962cr" event={"ID":"54a246a2-f674-4735-b295-b56699ece95b","Type":"ContainerStarted","Data":"81e3228ea3cb268d00e067cb20424ca316ff00a8dc2ce42ec5ac2ddb0165a7c2"} Jan 26 17:01:00 crc kubenswrapper[4856]: I0126 17:01:00.474596 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wvttb" event={"ID":"2d37efbf-d18f-486b-9b43-bc4d181af4ca","Type":"ContainerStarted","Data":"743ebe09ef635c21a62370a80c15b76e3ff5e7e1801bb955f28ed30f848dcca9"} Jan 26 17:01:00 crc kubenswrapper[4856]: I0126 17:01:00.475407 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-wvttb" Jan 26 17:01:00 crc kubenswrapper[4856]: I0126 17:01:00.476896 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rrhjv" event={"ID":"a1546392-4a69-4b12-8d7e-97450b73b7ca","Type":"ContainerStarted","Data":"6732b3a88a90b26d6f17e46917c6b34079eb5e0dff612eb539aca34c8e37ca6c"} Jan 26 17:01:00 crc kubenswrapper[4856]: I0126 17:01:00.478334 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8m4l6" event={"ID":"004316da-16cd-49ab-b14d-282c28da6fad","Type":"ContainerStarted","Data":"9c3a62568b04641da092b2d90197c2236c1d178b7784315aa24e6f7ee39206ab"} Jan 26 17:01:00 crc kubenswrapper[4856]: I0126 17:01:00.479413 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-zzxln" event={"ID":"129a0b30-7132-4e3c-ab84-208cae7cb2f2","Type":"ContainerStarted","Data":"c6ca70c17dd9f8d67fa1f3c69f2f9036d5c439cdefe6f738ef00b96ed19e3bb5"} Jan 26 17:01:00 crc kubenswrapper[4856]: I0126 17:01:00.480910 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:00 crc kubenswrapper[4856]: E0126 17:01:00.481646 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:00.98163303 +0000 UTC m=+156.934887011 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:00 crc kubenswrapper[4856]: I0126 17:01:00.482864 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-58fcz" event={"ID":"73de6ef2-e139-4185-9f56-9db885734ffe","Type":"ContainerStarted","Data":"1f14adac5cebdf92a2acc8ff852bdc7f82dfe245accd2081fd775ca85d151340"} Jan 26 17:01:00 crc kubenswrapper[4856]: I0126 17:01:00.601339 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:00 crc kubenswrapper[4856]: E0126 17:01:00.601655 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:01.101613378 +0000 UTC m=+157.054867359 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:00 crc kubenswrapper[4856]: I0126 17:01:00.603625 4856 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-wvttb container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" start-of-body= Jan 26 17:01:00 crc kubenswrapper[4856]: I0126 17:01:00.603655 4856 patch_prober.go:28] interesting pod/console-operator-58897d9998-4pbj2 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/readyz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Jan 26 17:01:00 crc kubenswrapper[4856]: I0126 17:01:00.603696 4856 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-k662z container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Jan 26 17:01:00 crc kubenswrapper[4856]: I0126 17:01:00.603738 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-4pbj2" podUID="831dc87e-8e14-43d3-a36e-dc7679041ae5" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/readyz\": dial tcp 10.217.0.13:8443: connect: connection refused" Jan 26 17:01:00 crc kubenswrapper[4856]: I0126 17:01:00.603766 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k662z" podUID="17a72e73-4d54-4a29-a85a-ecb1aff30d10" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" Jan 26 17:01:00 crc kubenswrapper[4856]: I0126 17:01:00.603671 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-wvttb" podUID="2d37efbf-d18f-486b-9b43-bc4d181af4ca" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" Jan 26 17:01:00 crc kubenswrapper[4856]: I0126 17:01:00.677156 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-8q6q4" event={"ID":"7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7","Type":"ContainerStarted","Data":"655c350d2621ac99cae47d6117abe996be96564e1734dccd0a74e6f8446d8e6d"} Jan 26 17:01:00 crc kubenswrapper[4856]: I0126 17:01:00.679163 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpqvc" event={"ID":"5fe6baed-ab97-4d8a-8be2-6f00f9698136","Type":"ContainerStarted","Data":"56133c3e036efeb9590dc043f9b9af766fce603e2c50cfdca46be37466b88f62"} Jan 26 17:01:00 crc kubenswrapper[4856]: I0126 17:01:00.679950 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpqvc" Jan 26 17:01:00 crc kubenswrapper[4856]: I0126 17:01:00.698371 4856 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-fpqvc container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Jan 26 17:01:00 crc kubenswrapper[4856]: I0126 17:01:00.698423 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpqvc" podUID="5fe6baed-ab97-4d8a-8be2-6f00f9698136" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Jan 26 17:01:00 crc kubenswrapper[4856]: I0126 17:01:00.700429 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-z7cgq" event={"ID":"113d2266-0e67-4e79-8a17-1a78cb9a13d5","Type":"ContainerStarted","Data":"4d819592d378e2681ee37d59708ad732f047e864429c81bd39834ba8340ed07b"} Jan 26 17:01:00 crc kubenswrapper[4856]: I0126 17:01:00.703223 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:00 crc kubenswrapper[4856]: E0126 17:01:00.705038 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:01.205018607 +0000 UTC m=+157.158272668 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:00 crc kubenswrapper[4856]: I0126 17:01:00.708122 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-h9b2g" event={"ID":"85f05bd5-ff83-4d29-9531-ab3499088095","Type":"ContainerStarted","Data":"a104ae01c9b404438ac5f68d31f78d35be23d15f7f55bc2db6e35373a2bd7220"} Jan 26 17:01:00 crc kubenswrapper[4856]: I0126 17:01:00.960733 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:00 crc kubenswrapper[4856]: E0126 17:01:00.962625 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:01.462580813 +0000 UTC m=+157.415834824 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:00 crc kubenswrapper[4856]: I0126 17:01:00.963466 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nn46h" event={"ID":"abbeffe1-cfd5-4476-9a8e-2ab5b4869444","Type":"ContainerStarted","Data":"c45386474fc1f7418b5e10777280d9281617bf3a100ed762dd0a20d022c0961d"} Jan 26 17:01:00 crc kubenswrapper[4856]: I0126 17:01:00.963583 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nn46h" Jan 26 17:01:00 crc kubenswrapper[4856]: I0126 17:01:00.965775 4856 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-nn46h container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" start-of-body= Jan 26 17:01:00 crc kubenswrapper[4856]: I0126 17:01:00.965834 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nn46h" podUID="abbeffe1-cfd5-4476-9a8e-2ab5b4869444" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" Jan 26 17:01:01 crc kubenswrapper[4856]: I0126 17:01:01.409948 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:01 crc kubenswrapper[4856]: I0126 17:01:01.410653 4856 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-h9b2g" Jan 26 17:01:01 crc kubenswrapper[4856]: E0126 17:01:01.412007 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:01.911990926 +0000 UTC m=+157.865244977 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:01 crc kubenswrapper[4856]: I0126 17:01:01.412784 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vmwvg" podStartSLOduration=126.412742588 podStartE2EDuration="2m6.412742588s" podCreationTimestamp="2026-01-26 16:58:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:01:01.053598217 +0000 UTC m=+157.006852218" watchObservedRunningTime="2026-01-26 17:01:01.412742588 +0000 UTC m=+157.365996569" Jan 26 17:01:01 crc kubenswrapper[4856]: I0126 17:01:01.418370 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-27vjc" podStartSLOduration=127.418350103 podStartE2EDuration="2m7.418350103s" podCreationTimestamp="2026-01-26 16:58:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:00:59.385542516 +0000 UTC m=+155.338796517" watchObservedRunningTime="2026-01-26 17:01:01.418350103 +0000 UTC m=+157.371604084" Jan 26 17:01:01 crc kubenswrapper[4856]: I0126 17:01:01.418903 4856 patch_prober.go:28] interesting pod/router-default-5444994796-h9b2g container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Jan 26 17:01:01 crc kubenswrapper[4856]: I0126 17:01:01.418998 4856 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9b2g" podUID="85f05bd5-ff83-4d29-9531-ab3499088095" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Jan 26 17:01:01 crc kubenswrapper[4856]: I0126 17:01:01.441722 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" event={"ID":"a6d331bd-2db3-4319-9f5c-db56d408d9e3","Type":"ContainerStarted","Data":"2e820399efdd76f35845a11fcb687223c7841ebdef9836a702a54e4f2ffafb8a"} Jan 26 17:01:01 crc kubenswrapper[4856]: I0126 17:01:01.444095 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qdmxz" event={"ID":"37a77f41-5dbf-4842-9e77-83dc22b50f4a","Type":"ContainerStarted","Data":"954083c25159407a3d66c9dc10b5da920dc4a8f6475aa2b216c7d3c0bc6df577"} Jan 26 17:01:01 crc kubenswrapper[4856]: I0126 17:01:01.448046 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-7l927" event={"ID":"94291fa4-24a5-499e-8143-89c8784d9284","Type":"ContainerStarted","Data":"d13c7142b05ed798c0e5b16508a221e2918021dbec60054995ac94f05ffdad09"} Jan 26 17:01:01 crc kubenswrapper[4856]: I0126 17:01:01.449823 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-7l927" Jan 26 17:01:01 crc kubenswrapper[4856]: I0126 17:01:01.449909 4856 patch_prober.go:28] interesting pod/downloads-7954f5f757-7l927 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Jan 26 17:01:01 crc kubenswrapper[4856]: I0126 17:01:01.450034 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-7l927" podUID="94291fa4-24a5-499e-8143-89c8784d9284" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Jan 26 17:01:01 crc kubenswrapper[4856]: I0126 17:01:01.480391 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-fbsj7" podStartSLOduration=10.480365942 podStartE2EDuration="10.480365942s" podCreationTimestamp="2026-01-26 17:00:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:01:01.478013432 +0000 UTC m=+157.431267433" watchObservedRunningTime="2026-01-26 17:01:01.480365942 +0000 UTC m=+157.433619923" Jan 26 17:01:01 crc kubenswrapper[4856]: I0126 17:01:01.511008 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:01 crc kubenswrapper[4856]: E0126 17:01:01.512372 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:02.012353445 +0000 UTC m=+157.965607426 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:01 crc kubenswrapper[4856]: I0126 17:01:01.519803 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-gz7kg" event={"ID":"033cb12f-278f-431a-8104-519db9a3152f","Type":"ContainerStarted","Data":"31c5ee65e1feb6ca373ddf1aafdd94c056851ed234ddbef669b778a0963b59f1"} Jan 26 17:01:01 crc kubenswrapper[4856]: I0126 17:01:01.530512 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2sfhr" event={"ID":"a4d83db5-776f-4e95-a6fa-b194344f9819","Type":"ContainerStarted","Data":"4547c83b8750bed5b224af4ece51940397e9dc8ec7129e201ab54486dd4fd6bf"} Jan 26 17:01:01 crc kubenswrapper[4856]: I0126 17:01:01.574883 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-cjzsq" event={"ID":"550752e4-a1d9-46f4-9118-9e9919b2fe6b","Type":"ContainerStarted","Data":"1a543371d727ed7dc1b3d683a835ad3e767732c10f87d4c8dddf2238ff49d0f2"} Jan 26 17:01:01 crc kubenswrapper[4856]: I0126 17:01:01.588214 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l9nqd" event={"ID":"05d74105-0ecd-41ac-9001-8b21b0fd6ba4","Type":"ContainerStarted","Data":"5eb5b3a9fa9c5e2805bb24a61b4fb04fdbb3fc2f8bdab7c1025e5e9a63ac14c0"} Jan 26 17:01:01 crc kubenswrapper[4856]: I0126 17:01:01.591813 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-ddghz" event={"ID":"beb6f283-75cb-4184-b985-4e6c095feca1","Type":"ContainerStarted","Data":"c4ae6024cd1a68b28c900e4e9366d73736bbb8b2a0126f7dde2b9f9ce32cbf08"} Jan 26 17:01:01 crc kubenswrapper[4856]: I0126 17:01:01.593512 4856 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-cb8nk container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" start-of-body= Jan 26 17:01:01 crc kubenswrapper[4856]: I0126 17:01:01.593571 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bjl7" Jan 26 17:01:01 crc kubenswrapper[4856]: I0126 17:01:01.593567 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" podUID="69008ed1-f3e5-400d-852f-adbcd94199f6" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" Jan 26 17:01:01 crc kubenswrapper[4856]: I0126 17:01:01.595684 4856 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-5bjl7 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Jan 26 17:01:01 crc kubenswrapper[4856]: I0126 17:01:01.595713 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bjl7" podUID="2ba3cf6a-a6be-4108-a155-c8bb530aa037" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Jan 26 17:01:01 crc kubenswrapper[4856]: I0126 17:01:01.615339 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:01 crc kubenswrapper[4856]: E0126 17:01:01.616446 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:02.116427734 +0000 UTC m=+158.069681795 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:01 crc kubenswrapper[4856]: I0126 17:01:01.633665 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-wvttb" podStartSLOduration=126.633643472 podStartE2EDuration="2m6.633643472s" podCreationTimestamp="2026-01-26 16:58:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:01:01.631494679 +0000 UTC m=+157.584748670" watchObservedRunningTime="2026-01-26 17:01:01.633643472 +0000 UTC m=+157.586897463" Jan 26 17:01:01 crc kubenswrapper[4856]: I0126 17:01:01.653611 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nn46h" podStartSLOduration=126.65359404 podStartE2EDuration="2m6.65359404s" podCreationTimestamp="2026-01-26 16:58:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:01:01.652483458 +0000 UTC m=+157.605737449" watchObservedRunningTime="2026-01-26 17:01:01.65359404 +0000 UTC m=+157.606848021" Jan 26 17:01:01 crc kubenswrapper[4856]: I0126 17:01:01.679006 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-h9b2g" podStartSLOduration=127.678984069 podStartE2EDuration="2m7.678984069s" podCreationTimestamp="2026-01-26 16:58:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:01:01.676561388 +0000 UTC m=+157.629815379" watchObservedRunningTime="2026-01-26 17:01:01.678984069 +0000 UTC m=+157.632238050" Jan 26 17:01:01 crc kubenswrapper[4856]: I0126 17:01:01.715979 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:01 crc kubenswrapper[4856]: E0126 17:01:01.717589 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:02.217568767 +0000 UTC m=+158.170822748 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:01 crc kubenswrapper[4856]: I0126 17:01:01.916591 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:01 crc kubenswrapper[4856]: I0126 17:01:01.992373 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-8q6q4" podStartSLOduration=61.99235393 podStartE2EDuration="1m1.99235393s" podCreationTimestamp="2026-01-26 17:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:01:01.744889933 +0000 UTC m=+157.698143924" watchObservedRunningTime="2026-01-26 17:01:01.99235393 +0000 UTC m=+157.945607911" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:02.740770 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:02.740825 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:02.740895 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:02.746620 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 17:01:04 crc kubenswrapper[4856]: E0126 17:01:02.742516 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:03.24247455 +0000 UTC m=+159.195728531 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:02.957683 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:02.958629 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:02.959613 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:02.743378 4856 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-5bjl7 container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:02.962195 4856 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bjl7" podUID="2ba3cf6a-a6be-4108-a155-c8bb530aa037" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:02.743760 4856 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-5bjl7 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:02.962302 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bjl7" podUID="2ba3cf6a-a6be-4108-a155-c8bb530aa037" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:02.744890 4856 patch_prober.go:28] interesting pod/router-default-5444994796-h9b2g container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:02.962348 4856 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9b2g" podUID="85f05bd5-ff83-4d29-9531-ab3499088095" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:02.964835 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:02.974370 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:02.976367 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:02.984977 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-962cr" podStartSLOduration=129.984956141 podStartE2EDuration="2m9.984956141s" podCreationTimestamp="2026-01-26 16:58:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:01:02.957208492 +0000 UTC m=+158.910462483" watchObservedRunningTime="2026-01-26 17:01:02.984956141 +0000 UTC m=+158.938210122" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:02.985278 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.010126 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.013910 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.027404 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.027697 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.035442 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpqvc" podStartSLOduration=128.035413499 podStartE2EDuration="2m8.035413499s" podCreationTimestamp="2026-01-26 16:58:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:01:03.009744172 +0000 UTC m=+158.962998163" watchObservedRunningTime="2026-01-26 17:01:03.035413499 +0000 UTC m=+158.988667490" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.056319 4856 patch_prober.go:28] interesting pod/router-default-5444994796-h9b2g container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.056354 4856 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9b2g" podUID="85f05bd5-ff83-4d29-9531-ab3499088095" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.058767 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:04 crc kubenswrapper[4856]: E0126 17:01:03.059087 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:03.559022345 +0000 UTC m=+159.512276316 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.063377 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-7xb2b" event={"ID":"77a97acb-2908-48fb-8bcd-0647f3e90160","Type":"ContainerStarted","Data":"a12bfae50af05a4fb853762d6e1356af54d9c96dca053f3311e468d2f354a5e2"} Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.066195 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qdmxz" event={"ID":"37a77f41-5dbf-4842-9e77-83dc22b50f4a","Type":"ContainerStarted","Data":"1982e8f56cf36af0e2f28ffd71f4bae06c0d07c81347f58eb49cd981b71ed717"} Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.094008 4856 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-nn46h container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" start-of-body= Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.094029 4856 patch_prober.go:28] interesting pod/console-operator-58897d9998-4pbj2 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/readyz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.094076 4856 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-wvttb container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" start-of-body= Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.094083 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nn46h" podUID="abbeffe1-cfd5-4476-9a8e-2ab5b4869444" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.094097 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-4pbj2" podUID="831dc87e-8e14-43d3-a36e-dc7679041ae5" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/readyz\": dial tcp 10.217.0.13:8443: connect: connection refused" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.094145 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-wvttb" podUID="2d37efbf-d18f-486b-9b43-bc4d181af4ca" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.094239 4856 patch_prober.go:28] interesting pod/downloads-7954f5f757-7l927 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.094257 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-7l927" podUID="94291fa4-24a5-499e-8143-89c8784d9284" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.094294 4856 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-fpqvc container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.094374 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpqvc" podUID="5fe6baed-ab97-4d8a-8be2-6f00f9698136" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.094658 4856 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-cb8nk container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" start-of-body= Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.094682 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" podUID="69008ed1-f3e5-400d-852f-adbcd94199f6" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.099995 4856 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-5bjl7 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.100066 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bjl7" podUID="2ba3cf6a-a6be-4108-a155-c8bb530aa037" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.100700 4856 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-k662z container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.100766 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k662z" podUID="17a72e73-4d54-4a29-a85a-ecb1aff30d10" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.167121 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.167196 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/56258114-1bee-4516-ab71-f60d15a9635d-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"56258114-1bee-4516-ab71-f60d15a9635d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.167353 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/56258114-1bee-4516-ab71-f60d15a9635d-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"56258114-1bee-4516-ab71-f60d15a9635d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 26 17:01:04 crc kubenswrapper[4856]: E0126 17:01:03.167796 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:03.667777192 +0000 UTC m=+159.621031173 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.176495 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rrhjv" podStartSLOduration=128.176465528 podStartE2EDuration="2m8.176465528s" podCreationTimestamp="2026-01-26 16:58:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:01:03.165726851 +0000 UTC m=+159.118980852" watchObservedRunningTime="2026-01-26 17:01:03.176465528 +0000 UTC m=+159.129719549" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.222764 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.224829 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-4pbj2" podStartSLOduration=129.224805974 podStartE2EDuration="2m9.224805974s" podCreationTimestamp="2026-01-26 16:58:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:01:03.223418243 +0000 UTC m=+159.176672234" watchObservedRunningTime="2026-01-26 17:01:03.224805974 +0000 UTC m=+159.178059955" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.306952 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.307335 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/56258114-1bee-4516-ab71-f60d15a9635d-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"56258114-1bee-4516-ab71-f60d15a9635d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.307602 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/56258114-1bee-4516-ab71-f60d15a9635d-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"56258114-1bee-4516-ab71-f60d15a9635d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 26 17:01:04 crc kubenswrapper[4856]: E0126 17:01:03.309157 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:03.80912035 +0000 UTC m=+159.762374331 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.315109 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/56258114-1bee-4516-ab71-f60d15a9635d-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"56258114-1bee-4516-ab71-f60d15a9635d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.354369 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k662z" podStartSLOduration=128.354344314 podStartE2EDuration="2m8.354344314s" podCreationTimestamp="2026-01-26 16:58:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:01:03.334407046 +0000 UTC m=+159.287661047" watchObservedRunningTime="2026-01-26 17:01:03.354344314 +0000 UTC m=+159.307598295" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.379479 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/56258114-1bee-4516-ab71-f60d15a9635d-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"56258114-1bee-4516-ab71-f60d15a9635d\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.408643 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:04 crc kubenswrapper[4856]: E0126 17:01:03.410515 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:03.91049865 +0000 UTC m=+159.863752631 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.431566 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-gz7kg" podStartSLOduration=128.43154245 podStartE2EDuration="2m8.43154245s" podCreationTimestamp="2026-01-26 16:58:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:01:03.430446328 +0000 UTC m=+159.383700319" watchObservedRunningTime="2026-01-26 17:01:03.43154245 +0000 UTC m=+159.384796441" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.669898 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:04 crc kubenswrapper[4856]: E0126 17:01:03.670351 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:04.170333912 +0000 UTC m=+160.123587893 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.673556 4856 patch_prober.go:28] interesting pod/console-f9d7485db-6qgnn container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.26:8443/health\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.673605 4856 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-6qgnn" podUID="b28404ed-2e71-4b3f-9140-35ee89dbc8f2" containerName="console" probeResult="failure" output="Get \"https://10.217.0.26:8443/health\": dial tcp 10.217.0.26:8443: connect: connection refused" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.707909 4856 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-lndnt container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.708023 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-lndnt" podUID="1afc0f4c-e02d-4a70-aaba-e761e8c04eee" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.775514 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:04 crc kubenswrapper[4856]: E0126 17:01:03.775981 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:04.275968638 +0000 UTC m=+160.229222619 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.777678 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-7xb2b" podStartSLOduration=128.777640217 podStartE2EDuration="2m8.777640217s" podCreationTimestamp="2026-01-26 16:58:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:01:03.753222607 +0000 UTC m=+159.706476588" watchObservedRunningTime="2026-01-26 17:01:03.777640217 +0000 UTC m=+159.730894198" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.812865 4856 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-fpqvc container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.812918 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpqvc" podUID="5fe6baed-ab97-4d8a-8be2-6f00f9698136" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.877929 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:04 crc kubenswrapper[4856]: E0126 17:01:03.878700 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:04.378679666 +0000 UTC m=+160.331933647 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.891988 4856 patch_prober.go:28] interesting pod/downloads-7954f5f757-7l927 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.892065 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-7l927" podUID="94291fa4-24a5-499e-8143-89c8784d9284" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.892158 4856 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-cb8nk container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" start-of-body= Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.892172 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" podUID="69008ed1-f3e5-400d-852f-adbcd94199f6" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.892248 4856 patch_prober.go:28] interesting pod/downloads-7954f5f757-7l927 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.892269 4856 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-7l927" podUID="94291fa4-24a5-499e-8143-89c8784d9284" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.895724 4856 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-mr7cp container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.19:5443/healthz\": dial tcp 10.217.0.19:5443: connect: connection refused" start-of-body= Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.895763 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mr7cp" podUID="f9b8f57e-00b9-4355-ace2-0319d320d208" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.19:5443/healthz\": dial tcp 10.217.0.19:5443: connect: connection refused" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.895831 4856 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-mr7cp container/packageserver namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.19:5443/healthz\": dial tcp 10.217.0.19:5443: connect: connection refused" start-of-body= Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.895845 4856 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mr7cp" podUID="f9b8f57e-00b9-4355-ace2-0319d320d208" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.19:5443/healthz\": dial tcp 10.217.0.19:5443: connect: connection refused" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.896385 4856 patch_prober.go:28] interesting pod/console-operator-58897d9998-4pbj2 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/readyz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.896407 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-4pbj2" podUID="831dc87e-8e14-43d3-a36e-dc7679041ae5" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/readyz\": dial tcp 10.217.0.13:8443: connect: connection refused" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.896451 4856 patch_prober.go:28] interesting pod/console-operator-58897d9998-4pbj2 container/console-operator namespace/openshift-console-operator: Liveness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.896467 4856 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console-operator/console-operator-58897d9998-4pbj2" podUID="831dc87e-8e14-43d3-a36e-dc7679041ae5" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.986451 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:04 crc kubenswrapper[4856]: E0126 17:01:03.987509 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:04.487444324 +0000 UTC m=+160.440698305 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:03.990384 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-7p5jt" podStartSLOduration=129.99035844 podStartE2EDuration="2m9.99035844s" podCreationTimestamp="2026-01-26 16:58:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:01:03.986218318 +0000 UTC m=+159.939472309" watchObservedRunningTime="2026-01-26 17:01:03.99035844 +0000 UTC m=+159.943612421" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.028244 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-cjzsq" podStartSLOduration=129.028204916 podStartE2EDuration="2m9.028204916s" podCreationTimestamp="2026-01-26 16:58:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:01:04.021410696 +0000 UTC m=+159.974664707" watchObservedRunningTime="2026-01-26 17:01:04.028204916 +0000 UTC m=+159.981458897" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.066902 4856 patch_prober.go:28] interesting pod/router-default-5444994796-h9b2g container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.067231 4856 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9b2g" podUID="85f05bd5-ff83-4d29-9531-ab3499088095" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.091563 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:04 crc kubenswrapper[4856]: E0126 17:01:04.091849 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:04.591830442 +0000 UTC m=+160.545084423 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.165482 4856 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-k662z container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.165556 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k662z" podUID="17a72e73-4d54-4a29-a85a-ecb1aff30d10" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.165472 4856 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-k662z container/olm-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.165605 4856 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k662z" podUID="17a72e73-4d54-4a29-a85a-ecb1aff30d10" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.165870 4856 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-nn46h container/catalog-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" start-of-body= Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.165891 4856 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nn46h" podUID="abbeffe1-cfd5-4476-9a8e-2ab5b4869444" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.165945 4856 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-nn46h container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" start-of-body= Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.166009 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nn46h" podUID="abbeffe1-cfd5-4476-9a8e-2ab5b4869444" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.206001 4856 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-wvttb container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" start-of-body= Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.206168 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-wvttb" podUID="2d37efbf-d18f-486b-9b43-bc4d181af4ca" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.206271 4856 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-wvttb container/marketplace-operator namespace/openshift-marketplace: Liveness probe status=failure output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" start-of-body= Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.206293 4856 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/marketplace-operator-79b997595-wvttb" podUID="2d37efbf-d18f-486b-9b43-bc4d181af4ca" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.207059 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:04 crc kubenswrapper[4856]: E0126 17:01:04.207454 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:04.707442511 +0000 UTC m=+160.660696502 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.308006 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:04 crc kubenswrapper[4856]: E0126 17:01:04.308251 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:04.808223033 +0000 UTC m=+160.761477014 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.308726 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:04 crc kubenswrapper[4856]: E0126 17:01:04.309159 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:04.80911513 +0000 UTC m=+160.762369111 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.317420 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-7l927" podStartSLOduration=130.317390754 podStartE2EDuration="2m10.317390754s" podCreationTimestamp="2026-01-26 16:58:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:01:04.27385271 +0000 UTC m=+160.227106701" watchObservedRunningTime="2026-01-26 17:01:04.317390754 +0000 UTC m=+160.270644735" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.317629 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-qdmxz" podStartSLOduration=129.317619951 podStartE2EDuration="2m9.317619951s" podCreationTimestamp="2026-01-26 16:58:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:01:04.069291668 +0000 UTC m=+160.022545679" watchObservedRunningTime="2026-01-26 17:01:04.317619951 +0000 UTC m=+160.270873942" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.355494 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-l9nqd" podStartSLOduration=130.355472097 podStartE2EDuration="2m10.355472097s" podCreationTimestamp="2026-01-26 16:58:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:01:04.325758371 +0000 UTC m=+160.279012362" watchObservedRunningTime="2026-01-26 17:01:04.355472097 +0000 UTC m=+160.308726078" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.419726 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:04 crc kubenswrapper[4856]: E0126 17:01:04.420311 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:04.920281348 +0000 UTC m=+160.873535339 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.421170 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:04 crc kubenswrapper[4856]: E0126 17:01:04.421978 4856 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.015s" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.422017 4856 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-6qgnn" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.422050 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-z7cgq" event={"ID":"113d2266-0e67-4e79-8a17-1a78cb9a13d5","Type":"ContainerStarted","Data":"c6c4d71b01ea7a92bb160f8285dd5ca4166d7aea7bf52c195a12c0adddc54878"} Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.422079 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-6qgnn" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.422097 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-zzxln" event={"ID":"129a0b30-7132-4e3c-ab84-208cae7cb2f2","Type":"ContainerStarted","Data":"aabc7c67d807fdf1f4ca5027d6328c13318343ee6b7d43d56a7f335230cc215a"} Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.422110 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mr7cp" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.422122 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-ddghz" event={"ID":"beb6f283-75cb-4184-b985-4e6c095feca1","Type":"ContainerStarted","Data":"6bdbb4d282933915ce00d8a66dcb4c0d6922b65c0b34b1d7a402740a53527a2b"} Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.422174 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-h9b2g" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.422197 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2sfhr" event={"ID":"a4d83db5-776f-4e95-a6fa-b194344f9819","Type":"ContainerStarted","Data":"4f799d007cd14db60504ef7895ca55c01499d25a3d916f2b08952f82e0ea032c"} Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.422212 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8m4l6" event={"ID":"004316da-16cd-49ab-b14d-282c28da6fad","Type":"ContainerStarted","Data":"57c6b7055d01d117511ba1b8338e885c578822b2d534f7a2171bfaf3838c3df7"} Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.422227 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8m4l6" Jan 26 17:01:04 crc kubenswrapper[4856]: E0126 17:01:04.422489 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:04.922474783 +0000 UTC m=+160.875728774 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.422712 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4w5bf" event={"ID":"0c1af7db-aa80-4cb0-a9cb-5afdf677f28c","Type":"ContainerStarted","Data":"d2920aad12d06efdb0eab5696d56fd3226d3268bbf899b8a38b443ca3cd7108f"} Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.423331 4856 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-k662z container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.423438 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k662z" podUID="17a72e73-4d54-4a29-a85a-ecb1aff30d10" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.423477 4856 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-fpqvc container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.423612 4856 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-mr7cp container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.19:5443/healthz\": dial tcp 10.217.0.19:5443: connect: connection refused" start-of-body= Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.423647 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mr7cp" podUID="f9b8f57e-00b9-4355-ace2-0319d320d208" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.19:5443/healthz\": dial tcp 10.217.0.19:5443: connect: connection refused" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.423726 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpqvc" podUID="5fe6baed-ab97-4d8a-8be2-6f00f9698136" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.423738 4856 patch_prober.go:28] interesting pod/console-operator-58897d9998-4pbj2 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/readyz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.423572 4856 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-wvttb container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" start-of-body= Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.423879 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-4pbj2" podUID="831dc87e-8e14-43d3-a36e-dc7679041ae5" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/readyz\": dial tcp 10.217.0.13:8443: connect: connection refused" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.423911 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-wvttb" podUID="2d37efbf-d18f-486b-9b43-bc4d181af4ca" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.424721 4856 patch_prober.go:28] interesting pod/downloads-7954f5f757-7l927 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.425061 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-7l927" podUID="94291fa4-24a5-499e-8143-89c8784d9284" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.530899 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:04 crc kubenswrapper[4856]: E0126 17:01:04.532122 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:05.032101056 +0000 UTC m=+160.985355037 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.637767 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:04 crc kubenswrapper[4856]: E0126 17:01:04.638504 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:05.138474422 +0000 UTC m=+161.091728443 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.639945 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bjl7" podStartSLOduration=130.639934035 podStartE2EDuration="2m10.639934035s" podCreationTimestamp="2026-01-26 16:58:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:01:04.493400294 +0000 UTC m=+160.446654295" watchObservedRunningTime="2026-01-26 17:01:04.639934035 +0000 UTC m=+160.593188036" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.642208 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-2sfhr" podStartSLOduration=129.642196672 podStartE2EDuration="2m9.642196672s" podCreationTimestamp="2026-01-26 16:58:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:01:04.63703063 +0000 UTC m=+160.590284621" watchObservedRunningTime="2026-01-26 17:01:04.642196672 +0000 UTC m=+160.595450663" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.759237 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:04 crc kubenswrapper[4856]: E0126 17:01:04.759736 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:05.259707548 +0000 UTC m=+161.212961529 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.867347 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:04 crc kubenswrapper[4856]: E0126 17:01:04.867753 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:05.367737003 +0000 UTC m=+161.320990984 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.950755 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8m4l6" podStartSLOduration=129.950729531 podStartE2EDuration="2m9.950729531s" podCreationTimestamp="2026-01-26 16:58:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:01:04.86628674 +0000 UTC m=+160.819540761" watchObservedRunningTime="2026-01-26 17:01:04.950729531 +0000 UTC m=+160.903983512" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.952831 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-ddghz" podStartSLOduration=129.952821092 podStartE2EDuration="2m9.952821092s" podCreationTimestamp="2026-01-26 16:58:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:01:04.925098445 +0000 UTC m=+160.878352446" watchObservedRunningTime="2026-01-26 17:01:04.952821092 +0000 UTC m=+160.906075073" Jan 26 17:01:04 crc kubenswrapper[4856]: I0126 17:01:04.970941 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:04 crc kubenswrapper[4856]: E0126 17:01:04.971358 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:05.471322488 +0000 UTC m=+161.424576459 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:05 crc kubenswrapper[4856]: I0126 17:01:05.056728 4856 patch_prober.go:28] interesting pod/router-default-5444994796-h9b2g container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Jan 26 17:01:05 crc kubenswrapper[4856]: I0126 17:01:05.056796 4856 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9b2g" podUID="85f05bd5-ff83-4d29-9531-ab3499088095" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Jan 26 17:01:05 crc kubenswrapper[4856]: I0126 17:01:05.064497 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 26 17:01:05 crc kubenswrapper[4856]: I0126 17:01:05.072962 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:05 crc kubenswrapper[4856]: E0126 17:01:05.073915 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:05.573898293 +0000 UTC m=+161.527152274 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:05 crc kubenswrapper[4856]: I0126 17:01:05.200191 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:05 crc kubenswrapper[4856]: E0126 17:01:05.201549 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:05.701511746 +0000 UTC m=+161.654765737 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:05 crc kubenswrapper[4856]: I0126 17:01:05.262171 4856 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-5bjl7 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Jan 26 17:01:05 crc kubenswrapper[4856]: I0126 17:01:05.262331 4856 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-5bjl7 container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Jan 26 17:01:05 crc kubenswrapper[4856]: I0126 17:01:05.262384 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bjl7" podUID="2ba3cf6a-a6be-4108-a155-c8bb530aa037" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Jan 26 17:01:05 crc kubenswrapper[4856]: I0126 17:01:05.262466 4856 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bjl7" podUID="2ba3cf6a-a6be-4108-a155-c8bb530aa037" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Jan 26 17:01:05 crc kubenswrapper[4856]: I0126 17:01:05.301831 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:05 crc kubenswrapper[4856]: E0126 17:01:05.302564 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:05.802490844 +0000 UTC m=+161.755744815 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:05 crc kubenswrapper[4856]: I0126 17:01:05.502393 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:05 crc kubenswrapper[4856]: E0126 17:01:05.505670 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:06.005640485 +0000 UTC m=+161.958894466 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:05 crc kubenswrapper[4856]: I0126 17:01:05.642388 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:05 crc kubenswrapper[4856]: E0126 17:01:05.647825 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:06.147808597 +0000 UTC m=+162.101062578 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:05 crc kubenswrapper[4856]: I0126 17:01:05.655841 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-4w5bf" podStartSLOduration=131.655811043 podStartE2EDuration="2m11.655811043s" podCreationTimestamp="2026-01-26 16:58:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:01:05.030816692 +0000 UTC m=+160.984070673" watchObservedRunningTime="2026-01-26 17:01:05.655811043 +0000 UTC m=+161.609065024" Jan 26 17:01:05 crc kubenswrapper[4856]: I0126 17:01:05.743123 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:05 crc kubenswrapper[4856]: E0126 17:01:05.743446 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:06.243414917 +0000 UTC m=+162.196668898 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:05 crc kubenswrapper[4856]: I0126 17:01:05.895030 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:05 crc kubenswrapper[4856]: E0126 17:01:05.895468 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:06.39545241 +0000 UTC m=+162.348706391 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:06 crc kubenswrapper[4856]: I0126 17:01:06.018606 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:06 crc kubenswrapper[4856]: E0126 17:01:06.018968 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:06.518951772 +0000 UTC m=+162.472205753 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:06 crc kubenswrapper[4856]: I0126 17:01:06.139629 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:06 crc kubenswrapper[4856]: E0126 17:01:06.140311 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:06.640295131 +0000 UTC m=+162.593549122 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:06 crc kubenswrapper[4856]: I0126 17:01:06.241136 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:06 crc kubenswrapper[4856]: E0126 17:01:06.241388 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:06.74137154 +0000 UTC m=+162.694625521 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:06 crc kubenswrapper[4856]: I0126 17:01:06.344342 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:06 crc kubenswrapper[4856]: E0126 17:01:06.344732 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:06.844716888 +0000 UTC m=+162.797970869 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:06 crc kubenswrapper[4856]: I0126 17:01:06.444304 4856 patch_prober.go:28] interesting pod/router-default-5444994796-h9b2g container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Jan 26 17:01:06 crc kubenswrapper[4856]: I0126 17:01:06.444443 4856 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9b2g" podUID="85f05bd5-ff83-4d29-9531-ab3499088095" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Jan 26 17:01:06 crc kubenswrapper[4856]: I0126 17:01:06.444884 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:06 crc kubenswrapper[4856]: E0126 17:01:06.445713 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:06.945697836 +0000 UTC m=+162.898951807 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:06 crc kubenswrapper[4856]: I0126 17:01:06.547664 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:06 crc kubenswrapper[4856]: E0126 17:01:06.548305 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:07.048290771 +0000 UTC m=+163.001544752 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:06 crc kubenswrapper[4856]: I0126 17:01:06.699494 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:06 crc kubenswrapper[4856]: E0126 17:01:06.700105 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:07.200089288 +0000 UTC m=+163.153343269 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:06 crc kubenswrapper[4856]: I0126 17:01:06.712263 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-58fcz" event={"ID":"73de6ef2-e139-4185-9f56-9db885734ffe","Type":"ContainerStarted","Data":"c6d018030b40179bdc7061fd342e5778334c8d595e566f8dccd7909d4cdfdb65"} Jan 26 17:01:06 crc kubenswrapper[4856]: I0126 17:01:06.742251 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-dgcqn" event={"ID":"ac10f013-cd1f-47e0-8f1c-5ff4e6e75784","Type":"ContainerStarted","Data":"721b13b3a9452dc526cb9f1d2ec4056ad4445955462616f9a8c3ab38495dfb4a"} Jan 26 17:01:06 crc kubenswrapper[4856]: I0126 17:01:06.743044 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-dgcqn" Jan 26 17:01:06 crc kubenswrapper[4856]: I0126 17:01:06.743324 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"7478646e9933e69628e5ab6af89d35504a3abd1eb3586313caed673dc6a2653e"} Jan 26 17:01:06 crc kubenswrapper[4856]: I0126 17:01:06.744746 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-vfm8t" event={"ID":"c8657575-cd22-4ebc-ae9d-4174366985d3","Type":"ContainerStarted","Data":"386ffe7f255acb85e646366b8112901fa773aedab0c0512515eb85e39d1d12ef"} Jan 26 17:01:06 crc kubenswrapper[4856]: I0126 17:01:06.746832 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" event={"ID":"a6d331bd-2db3-4319-9f5c-db56d408d9e3","Type":"ContainerStarted","Data":"4d58f3b5a90464745c269a6b00df4216969791b4eff2dcd4354fd144525f157c"} Jan 26 17:01:06 crc kubenswrapper[4856]: I0126 17:01:06.818949 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6cghs" event={"ID":"81c2f96b-55e0-483b-b72c-df7e156e9218","Type":"ContainerStarted","Data":"656e05b6881c7f1686e0f11444cdd88478ced84f0212c2e6645b5e4be4f15871"} Jan 26 17:01:06 crc kubenswrapper[4856]: I0126 17:01:06.822776 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:06 crc kubenswrapper[4856]: E0126 17:01:06.828134 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:07.328114723 +0000 UTC m=+163.281368704 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:07 crc kubenswrapper[4856]: I0126 17:01:07.195424 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:07 crc kubenswrapper[4856]: E0126 17:01:07.196956 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:07.69692969 +0000 UTC m=+163.650183721 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:07 crc kubenswrapper[4856]: I0126 17:01:07.199693 4856 patch_prober.go:28] interesting pod/router-default-5444994796-h9b2g container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Jan 26 17:01:07 crc kubenswrapper[4856]: I0126 17:01:07.199753 4856 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9b2g" podUID="85f05bd5-ff83-4d29-9531-ab3499088095" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Jan 26 17:01:07 crc kubenswrapper[4856]: I0126 17:01:07.305495 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:07 crc kubenswrapper[4856]: E0126 17:01:07.306165 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:07.8061503 +0000 UTC m=+163.759404281 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:07 crc kubenswrapper[4856]: I0126 17:01:07.413685 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:07 crc kubenswrapper[4856]: E0126 17:01:07.414353 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:07.91431132 +0000 UTC m=+163.867565311 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:07 crc kubenswrapper[4856]: I0126 17:01:07.414793 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:07 crc kubenswrapper[4856]: E0126 17:01:07.415543 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:07.915514595 +0000 UTC m=+163.868768576 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:07 crc kubenswrapper[4856]: I0126 17:01:07.518116 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-z7cgq" podStartSLOduration=133.518097931 podStartE2EDuration="2m13.518097931s" podCreationTimestamp="2026-01-26 16:58:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:01:07.517129372 +0000 UTC m=+163.470383353" watchObservedRunningTime="2026-01-26 17:01:07.518097931 +0000 UTC m=+163.471351912" Jan 26 17:01:07 crc kubenswrapper[4856]: I0126 17:01:07.524021 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:07 crc kubenswrapper[4856]: E0126 17:01:07.524299 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:08.024256432 +0000 UTC m=+163.977510413 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:07 crc kubenswrapper[4856]: I0126 17:01:07.525002 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:07 crc kubenswrapper[4856]: E0126 17:01:07.525557 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:08.02554286 +0000 UTC m=+163.978796841 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:07 crc kubenswrapper[4856]: I0126 17:01:07.559278 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 26 17:01:07 crc kubenswrapper[4856]: I0126 17:01:07.560022 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 26 17:01:07 crc kubenswrapper[4856]: I0126 17:01:07.622753 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Jan 26 17:01:07 crc kubenswrapper[4856]: I0126 17:01:07.622931 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Jan 26 17:01:07 crc kubenswrapper[4856]: I0126 17:01:07.626663 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:07 crc kubenswrapper[4856]: I0126 17:01:07.626985 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d8c22047-144c-402a-80c5-c206539b6826-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"d8c22047-144c-402a-80c5-c206539b6826\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 26 17:01:07 crc kubenswrapper[4856]: I0126 17:01:07.627095 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d8c22047-144c-402a-80c5-c206539b6826-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"d8c22047-144c-402a-80c5-c206539b6826\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 26 17:01:07 crc kubenswrapper[4856]: E0126 17:01:07.627288 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:08.12726609 +0000 UTC m=+164.080520071 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:07 crc kubenswrapper[4856]: I0126 17:01:07.726928 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 26 17:01:07 crc kubenswrapper[4856]: I0126 17:01:07.727665 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-zzxln" podStartSLOduration=132.72764835 podStartE2EDuration="2m12.72764835s" podCreationTimestamp="2026-01-26 16:58:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:01:07.704941751 +0000 UTC m=+163.658195732" watchObservedRunningTime="2026-01-26 17:01:07.72764835 +0000 UTC m=+163.680902331" Jan 26 17:01:07 crc kubenswrapper[4856]: I0126 17:01:07.728312 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d8c22047-144c-402a-80c5-c206539b6826-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"d8c22047-144c-402a-80c5-c206539b6826\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 26 17:01:07 crc kubenswrapper[4856]: I0126 17:01:07.738852 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:07 crc kubenswrapper[4856]: I0126 17:01:07.738898 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d8c22047-144c-402a-80c5-c206539b6826-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"d8c22047-144c-402a-80c5-c206539b6826\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 26 17:01:07 crc kubenswrapper[4856]: I0126 17:01:07.739041 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d8c22047-144c-402a-80c5-c206539b6826-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"d8c22047-144c-402a-80c5-c206539b6826\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 26 17:01:07 crc kubenswrapper[4856]: E0126 17:01:07.739388 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:08.239372556 +0000 UTC m=+164.192626537 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:07 crc kubenswrapper[4856]: I0126 17:01:07.856047 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:07 crc kubenswrapper[4856]: E0126 17:01:07.856912 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:08.356893512 +0000 UTC m=+164.310147493 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:07 crc kubenswrapper[4856]: I0126 17:01:07.896893 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d8c22047-144c-402a-80c5-c206539b6826-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"d8c22047-144c-402a-80c5-c206539b6826\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 26 17:01:08 crc kubenswrapper[4856]: I0126 17:01:07.964706 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:08 crc kubenswrapper[4856]: E0126 17:01:07.965162 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:08.465147614 +0000 UTC m=+164.418401595 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:08 crc kubenswrapper[4856]: I0126 17:01:08.071502 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:08 crc kubenswrapper[4856]: E0126 17:01:08.072199 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:08.5721716 +0000 UTC m=+164.525425581 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:08 crc kubenswrapper[4856]: I0126 17:01:08.081758 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 26 17:01:08 crc kubenswrapper[4856]: I0126 17:01:08.111733 4856 patch_prober.go:28] interesting pod/router-default-5444994796-h9b2g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 17:01:08 crc kubenswrapper[4856]: [-]has-synced failed: reason withheld Jan 26 17:01:08 crc kubenswrapper[4856]: [+]process-running ok Jan 26 17:01:08 crc kubenswrapper[4856]: healthz check failed Jan 26 17:01:08 crc kubenswrapper[4856]: I0126 17:01:08.111823 4856 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9b2g" podUID="85f05bd5-ff83-4d29-9531-ab3499088095" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 17:01:08 crc kubenswrapper[4856]: I0126 17:01:08.206233 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:08 crc kubenswrapper[4856]: E0126 17:01:08.206676 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:08.706663936 +0000 UTC m=+164.659917917 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:08 crc kubenswrapper[4856]: I0126 17:01:08.307305 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:08 crc kubenswrapper[4856]: E0126 17:01:08.307437 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:08.807413117 +0000 UTC m=+164.760667098 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:08 crc kubenswrapper[4856]: I0126 17:01:08.307537 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:08 crc kubenswrapper[4856]: E0126 17:01:08.307987 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:08.807970904 +0000 UTC m=+164.761224885 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:08 crc kubenswrapper[4856]: I0126 17:01:08.309553 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"0b843bfa4e4f72e7f44348fbd591ae8f8a30a66e1fcd865ab9fae8f878ae08fa"} Jan 26 17:01:08 crc kubenswrapper[4856]: I0126 17:01:08.319911 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"69d2574f87ca003643e7ca59ab73844ae16eabe1203ec86211d4b7f831a4c54e"} Jan 26 17:01:08 crc kubenswrapper[4856]: I0126 17:01:08.320308 4856 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-5bjl7 container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Jan 26 17:01:08 crc kubenswrapper[4856]: I0126 17:01:08.320347 4856 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bjl7" podUID="2ba3cf6a-a6be-4108-a155-c8bb530aa037" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Jan 26 17:01:08 crc kubenswrapper[4856]: I0126 17:01:08.320381 4856 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bjl7" Jan 26 17:01:08 crc kubenswrapper[4856]: I0126 17:01:08.320937 4856 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="openshift-config-operator" containerStatusID={"Type":"cri-o","ID":"7696af9bf5eb7a27c45bc9a500fea17921f66464546a3b193d1abfd56ccd50c4"} pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bjl7" containerMessage="Container openshift-config-operator failed liveness probe, will be restarted" Jan 26 17:01:08 crc kubenswrapper[4856]: I0126 17:01:08.321134 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bjl7" podUID="2ba3cf6a-a6be-4108-a155-c8bb530aa037" containerName="openshift-config-operator" containerID="cri-o://7696af9bf5eb7a27c45bc9a500fea17921f66464546a3b193d1abfd56ccd50c4" gracePeriod=30 Jan 26 17:01:08 crc kubenswrapper[4856]: I0126 17:01:08.321340 4856 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-5bjl7 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Jan 26 17:01:08 crc kubenswrapper[4856]: I0126 17:01:08.321382 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bjl7" podUID="2ba3cf6a-a6be-4108-a155-c8bb530aa037" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Jan 26 17:01:08 crc kubenswrapper[4856]: I0126 17:01:08.321782 4856 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-5bjl7 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Jan 26 17:01:08 crc kubenswrapper[4856]: I0126 17:01:08.321803 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bjl7" podUID="2ba3cf6a-a6be-4108-a155-c8bb530aa037" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Jan 26 17:01:08 crc kubenswrapper[4856]: I0126 17:01:08.378752 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-dgcqn" podStartSLOduration=17.3787288 podStartE2EDuration="17.3787288s" podCreationTimestamp="2026-01-26 17:00:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:01:08.10641412 +0000 UTC m=+164.059668111" watchObservedRunningTime="2026-01-26 17:01:08.3787288 +0000 UTC m=+164.331982781" Jan 26 17:01:08 crc kubenswrapper[4856]: I0126 17:01:08.433772 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"108b2ba5b94983d79d93282a417b0fe8f3ba567db5dc09cf99793744d5af7e2f"} Jan 26 17:01:08 crc kubenswrapper[4856]: I0126 17:01:08.435626 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:08 crc kubenswrapper[4856]: I0126 17:01:08.453314 4856 patch_prober.go:28] interesting pod/dns-default-dgcqn container/dns namespace/openshift-dns: Readiness probe status=failure output="Get \"http://10.217.0.43:8181/ready\": dial tcp 10.217.0.43:8181: connect: connection refused" start-of-body= Jan 26 17:01:08 crc kubenswrapper[4856]: I0126 17:01:08.453424 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-dns/dns-default-dgcqn" podUID="ac10f013-cd1f-47e0-8f1c-5ff4e6e75784" containerName="dns" probeResult="failure" output="Get \"http://10.217.0.43:8181/ready\": dial tcp 10.217.0.43:8181: connect: connection refused" Jan 26 17:01:08 crc kubenswrapper[4856]: E0126 17:01:08.506428 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:09.006398845 +0000 UTC m=+164.959652826 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:08 crc kubenswrapper[4856]: I0126 17:01:08.603497 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:08 crc kubenswrapper[4856]: E0126 17:01:08.606372 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:09.106353793 +0000 UTC m=+165.059607764 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:08 crc kubenswrapper[4856]: I0126 17:01:08.699712 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" Jan 26 17:01:08 crc kubenswrapper[4856]: I0126 17:01:08.700122 4856 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" Jan 26 17:01:08 crc kubenswrapper[4856]: I0126 17:01:08.704409 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:08 crc kubenswrapper[4856]: E0126 17:01:08.704944 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:09.20492577 +0000 UTC m=+165.158179751 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:08 crc kubenswrapper[4856]: I0126 17:01:08.723733 4856 patch_prober.go:28] interesting pod/apiserver-76f77b778f-6rlxp container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.5:8443/livez\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Jan 26 17:01:08 crc kubenswrapper[4856]: I0126 17:01:08.723818 4856 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" podUID="a6d331bd-2db3-4319-9f5c-db56d408d9e3" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.5:8443/livez\": dial tcp 10.217.0.5:8443: connect: connection refused" Jan 26 17:01:08 crc kubenswrapper[4856]: I0126 17:01:08.807224 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:08 crc kubenswrapper[4856]: E0126 17:01:08.807597 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:09.307583617 +0000 UTC m=+165.260837598 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:08 crc kubenswrapper[4856]: I0126 17:01:08.907041 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-58fcz" podStartSLOduration=134.907016319 podStartE2EDuration="2m14.907016319s" podCreationTimestamp="2026-01-26 16:58:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:01:08.381822242 +0000 UTC m=+164.335076223" watchObservedRunningTime="2026-01-26 17:01:08.907016319 +0000 UTC m=+164.860270300" Jan 26 17:01:08 crc kubenswrapper[4856]: I0126 17:01:08.908287 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:08 crc kubenswrapper[4856]: E0126 17:01:08.908735 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:09.408717719 +0000 UTC m=+165.361971700 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:08 crc kubenswrapper[4856]: I0126 17:01:08.908759 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 26 17:01:09 crc kubenswrapper[4856]: I0126 17:01:09.017351 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:09 crc kubenswrapper[4856]: E0126 17:01:09.017904 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:09.517891789 +0000 UTC m=+165.471145770 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:09 crc kubenswrapper[4856]: I0126 17:01:09.094633 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6cghs" Jan 26 17:01:09 crc kubenswrapper[4856]: I0126 17:01:09.094998 4856 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6cghs" Jan 26 17:01:09 crc kubenswrapper[4856]: I0126 17:01:09.103917 4856 patch_prober.go:28] interesting pod/router-default-5444994796-h9b2g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 17:01:09 crc kubenswrapper[4856]: [-]has-synced failed: reason withheld Jan 26 17:01:09 crc kubenswrapper[4856]: [+]process-running ok Jan 26 17:01:09 crc kubenswrapper[4856]: healthz check failed Jan 26 17:01:09 crc kubenswrapper[4856]: I0126 17:01:09.103983 4856 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9b2g" podUID="85f05bd5-ff83-4d29-9531-ab3499088095" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 17:01:09 crc kubenswrapper[4856]: I0126 17:01:09.115832 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-dgcqn" Jan 26 17:01:09 crc kubenswrapper[4856]: I0126 17:01:09.121102 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:09 crc kubenswrapper[4856]: E0126 17:01:09.121388 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:09.621372021 +0000 UTC m=+165.574626002 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:09 crc kubenswrapper[4856]: I0126 17:01:09.121816 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:09 crc kubenswrapper[4856]: E0126 17:01:09.122128 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:09.622121833 +0000 UTC m=+165.575375804 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:09 crc kubenswrapper[4856]: I0126 17:01:09.209100 4856 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-6cghs container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.22:8443/livez\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Jan 26 17:01:09 crc kubenswrapper[4856]: I0126 17:01:09.209156 4856 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6cghs" podUID="81c2f96b-55e0-483b-b72c-df7e156e9218" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.22:8443/livez\": dial tcp 10.217.0.22:8443: connect: connection refused" Jan 26 17:01:09 crc kubenswrapper[4856]: I0126 17:01:09.222745 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:09 crc kubenswrapper[4856]: E0126 17:01:09.223340 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:09.723309997 +0000 UTC m=+165.676563978 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:09 crc kubenswrapper[4856]: I0126 17:01:09.324417 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:09 crc kubenswrapper[4856]: E0126 17:01:09.325394 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:09.825364006 +0000 UTC m=+165.778617987 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:09 crc kubenswrapper[4856]: I0126 17:01:09.397293 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6cghs" podStartSLOduration=134.397269197 podStartE2EDuration="2m14.397269197s" podCreationTimestamp="2026-01-26 16:58:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:01:09.115494517 +0000 UTC m=+165.068748518" watchObservedRunningTime="2026-01-26 17:01:09.397269197 +0000 UTC m=+165.350523178" Jan 26 17:01:09 crc kubenswrapper[4856]: I0126 17:01:09.429272 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:09 crc kubenswrapper[4856]: E0126 17:01:09.429981 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:09.929961401 +0000 UTC m=+165.883215382 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:09 crc kubenswrapper[4856]: I0126 17:01:09.450428 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-config-operator_openshift-config-operator-7777fb866f-5bjl7_2ba3cf6a-a6be-4108-a155-c8bb530aa037/openshift-config-operator/0.log" Jan 26 17:01:09 crc kubenswrapper[4856]: I0126 17:01:09.452234 4856 generic.go:334] "Generic (PLEG): container finished" podID="2ba3cf6a-a6be-4108-a155-c8bb530aa037" containerID="7696af9bf5eb7a27c45bc9a500fea17921f66464546a3b193d1abfd56ccd50c4" exitCode=2 Jan 26 17:01:09 crc kubenswrapper[4856]: I0126 17:01:09.531005 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:09 crc kubenswrapper[4856]: E0126 17:01:09.534022 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:10.034002369 +0000 UTC m=+165.987256360 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:09 crc kubenswrapper[4856]: I0126 17:01:09.566753 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 17:01:09 crc kubenswrapper[4856]: I0126 17:01:09.566790 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"02669c91c1c07af72d4562227d525afec05b92f59598d7c5e3d7bae5dd7ee11d"} Jan 26 17:01:09 crc kubenswrapper[4856]: I0126 17:01:09.566803 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bjl7" event={"ID":"2ba3cf6a-a6be-4108-a155-c8bb530aa037","Type":"ContainerDied","Data":"7696af9bf5eb7a27c45bc9a500fea17921f66464546a3b193d1abfd56ccd50c4"} Jan 26 17:01:09 crc kubenswrapper[4856]: I0126 17:01:09.566815 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"56258114-1bee-4516-ab71-f60d15a9635d","Type":"ContainerStarted","Data":"02721e310b92647ac526a2662cb5ceecb1c039ddbbb08ec7de0c0bb07775f5b7"} Jan 26 17:01:09 crc kubenswrapper[4856]: I0126 17:01:09.567299 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" podStartSLOduration=136.56727413 podStartE2EDuration="2m16.56727413s" podCreationTimestamp="2026-01-26 16:58:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:01:09.399296546 +0000 UTC m=+165.352550547" watchObservedRunningTime="2026-01-26 17:01:09.56727413 +0000 UTC m=+165.520528111" Jan 26 17:01:09 crc kubenswrapper[4856]: I0126 17:01:09.636311 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:09 crc kubenswrapper[4856]: E0126 17:01:09.636793 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:10.136775939 +0000 UTC m=+166.090029920 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:09 crc kubenswrapper[4856]: I0126 17:01:09.742436 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:09 crc kubenswrapper[4856]: E0126 17:01:09.742856 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:10.242840387 +0000 UTC m=+166.196094378 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:09 crc kubenswrapper[4856]: I0126 17:01:09.855776 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:09 crc kubenswrapper[4856]: E0126 17:01:09.856178 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:10.356141738 +0000 UTC m=+166.309395719 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:09 crc kubenswrapper[4856]: I0126 17:01:09.957981 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:09 crc kubenswrapper[4856]: E0126 17:01:09.958377 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:10.458362942 +0000 UTC m=+166.411616923 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:10 crc kubenswrapper[4856]: I0126 17:01:10.268741 4856 patch_prober.go:28] interesting pod/router-default-5444994796-h9b2g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 17:01:10 crc kubenswrapper[4856]: [-]has-synced failed: reason withheld Jan 26 17:01:10 crc kubenswrapper[4856]: [+]process-running ok Jan 26 17:01:10 crc kubenswrapper[4856]: healthz check failed Jan 26 17:01:10 crc kubenswrapper[4856]: I0126 17:01:10.268809 4856 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9b2g" podUID="85f05bd5-ff83-4d29-9531-ab3499088095" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 17:01:10 crc kubenswrapper[4856]: I0126 17:01:10.269610 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:10 crc kubenswrapper[4856]: E0126 17:01:10.269953 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:10.76993658 +0000 UTC m=+166.723190561 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:10 crc kubenswrapper[4856]: I0126 17:01:10.383781 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:10 crc kubenswrapper[4856]: E0126 17:01:10.384140 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:10.884127358 +0000 UTC m=+166.837381339 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:10 crc kubenswrapper[4856]: I0126 17:01:10.504120 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:10 crc kubenswrapper[4856]: E0126 17:01:10.504383 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:11.004367253 +0000 UTC m=+166.957621234 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:10 crc kubenswrapper[4856]: I0126 17:01:10.556999 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"fbce6e92307ee43d8be105237704a8ad799999f886a73ccdca4b10e58a820780"} Jan 26 17:01:10 crc kubenswrapper[4856]: I0126 17:01:10.630269 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:10 crc kubenswrapper[4856]: E0126 17:01:10.631121 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:11.131105511 +0000 UTC m=+167.084359492 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:10 crc kubenswrapper[4856]: I0126 17:01:10.886274 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:10 crc kubenswrapper[4856]: E0126 17:01:10.886873 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:11.386853393 +0000 UTC m=+167.340107374 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:10 crc kubenswrapper[4856]: I0126 17:01:10.993733 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:10 crc kubenswrapper[4856]: E0126 17:01:10.994124 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:11.494107246 +0000 UTC m=+167.447361237 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:11 crc kubenswrapper[4856]: I0126 17:01:11.556909 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:11 crc kubenswrapper[4856]: I0126 17:01:11.557399 4856 patch_prober.go:28] interesting pod/router-default-5444994796-h9b2g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 17:01:11 crc kubenswrapper[4856]: [-]has-synced failed: reason withheld Jan 26 17:01:11 crc kubenswrapper[4856]: [+]process-running ok Jan 26 17:01:11 crc kubenswrapper[4856]: healthz check failed Jan 26 17:01:11 crc kubenswrapper[4856]: I0126 17:01:11.557479 4856 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9b2g" podUID="85f05bd5-ff83-4d29-9531-ab3499088095" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 17:01:11 crc kubenswrapper[4856]: I0126 17:01:11.558376 4856 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-5bjl7 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Jan 26 17:01:11 crc kubenswrapper[4856]: I0126 17:01:11.558452 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bjl7" podUID="2ba3cf6a-a6be-4108-a155-c8bb530aa037" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Jan 26 17:01:11 crc kubenswrapper[4856]: E0126 17:01:11.558644 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:12.558621673 +0000 UTC m=+168.511875674 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:11 crc kubenswrapper[4856]: I0126 17:01:11.917744 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:12 crc kubenswrapper[4856]: E0126 17:01:11.918509 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:12.418484935 +0000 UTC m=+168.371738916 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:12 crc kubenswrapper[4856]: I0126 17:01:11.923414 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:12 crc kubenswrapper[4856]: E0126 17:01:11.941073 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:12.441051021 +0000 UTC m=+168.394305002 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:12 crc kubenswrapper[4856]: I0126 17:01:12.069100 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:12 crc kubenswrapper[4856]: I0126 17:01:12.069682 4856 patch_prober.go:28] interesting pod/router-default-5444994796-h9b2g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 17:01:12 crc kubenswrapper[4856]: [-]has-synced failed: reason withheld Jan 26 17:01:12 crc kubenswrapper[4856]: [+]process-running ok Jan 26 17:01:12 crc kubenswrapper[4856]: healthz check failed Jan 26 17:01:12 crc kubenswrapper[4856]: I0126 17:01:12.069733 4856 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9b2g" podUID="85f05bd5-ff83-4d29-9531-ab3499088095" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 17:01:12 crc kubenswrapper[4856]: E0126 17:01:12.069983 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:12.569955322 +0000 UTC m=+168.523209313 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:12 crc kubenswrapper[4856]: I0126 17:01:12.223294 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:12 crc kubenswrapper[4856]: E0126 17:01:12.223653 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:12.723641374 +0000 UTC m=+168.676895355 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:12 crc kubenswrapper[4856]: I0126 17:01:12.350036 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:12 crc kubenswrapper[4856]: E0126 17:01:12.350541 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:12.850505865 +0000 UTC m=+168.803759846 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:12 crc kubenswrapper[4856]: I0126 17:01:12.505130 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:12 crc kubenswrapper[4856]: E0126 17:01:12.505430 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:13.005418344 +0000 UTC m=+168.958672325 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:12 crc kubenswrapper[4856]: I0126 17:01:12.816806 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:12 crc kubenswrapper[4856]: E0126 17:01:12.817418 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:13.317402324 +0000 UTC m=+169.270656305 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:13 crc kubenswrapper[4856]: I0126 17:01:12.846032 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-config-operator_openshift-config-operator-7777fb866f-5bjl7_2ba3cf6a-a6be-4108-a155-c8bb530aa037/openshift-config-operator/0.log" Jan 26 17:01:13 crc kubenswrapper[4856]: I0126 17:01:12.850250 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bjl7" event={"ID":"2ba3cf6a-a6be-4108-a155-c8bb530aa037","Type":"ContainerStarted","Data":"6f3704a0f6342f24993fae1a0b1a22eff40cbcbc5b7f46e53e9ada0cdde1aa55"} Jan 26 17:01:13 crc kubenswrapper[4856]: I0126 17:01:12.851395 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bjl7" Jan 26 17:01:13 crc kubenswrapper[4856]: I0126 17:01:12.920650 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:13 crc kubenswrapper[4856]: E0126 17:01:12.921189 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:13.421165724 +0000 UTC m=+169.374419705 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:13 crc kubenswrapper[4856]: I0126 17:01:13.092722 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:13 crc kubenswrapper[4856]: I0126 17:01:13.093056 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"56258114-1bee-4516-ab71-f60d15a9635d","Type":"ContainerStarted","Data":"688d67199a7c309449abcb4f65cea024a314835283d336314d00644e67f44daf"} Jan 26 17:01:13 crc kubenswrapper[4856]: E0126 17:01:13.093161 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:13.593123485 +0000 UTC m=+169.546377466 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:13 crc kubenswrapper[4856]: I0126 17:01:13.110599 4856 patch_prober.go:28] interesting pod/router-default-5444994796-h9b2g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 17:01:13 crc kubenswrapper[4856]: [-]has-synced failed: reason withheld Jan 26 17:01:13 crc kubenswrapper[4856]: [+]process-running ok Jan 26 17:01:13 crc kubenswrapper[4856]: healthz check failed Jan 26 17:01:13 crc kubenswrapper[4856]: I0126 17:01:13.110686 4856 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9b2g" podUID="85f05bd5-ff83-4d29-9531-ab3499088095" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 17:01:13 crc kubenswrapper[4856]: I0126 17:01:13.635164 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:13 crc kubenswrapper[4856]: E0126 17:01:13.635921 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:14.63589754 +0000 UTC m=+170.589151521 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:13 crc kubenswrapper[4856]: I0126 17:01:13.655256 4856 patch_prober.go:28] interesting pod/console-f9d7485db-6qgnn container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.26:8443/health\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Jan 26 17:01:13 crc kubenswrapper[4856]: I0126 17:01:13.655321 4856 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-6qgnn" podUID="b28404ed-2e71-4b3f-9140-35ee89dbc8f2" containerName="console" probeResult="failure" output="Get \"https://10.217.0.26:8443/health\": dial tcp 10.217.0.26:8443: connect: connection refused" Jan 26 17:01:14 crc kubenswrapper[4856]: I0126 17:01:14.131374 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:14 crc kubenswrapper[4856]: E0126 17:01:14.131729 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:14.631717461 +0000 UTC m=+170.584971442 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:14 crc kubenswrapper[4856]: I0126 17:01:14.161858 4856 patch_prober.go:28] interesting pod/router-default-5444994796-h9b2g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 17:01:14 crc kubenswrapper[4856]: [-]has-synced failed: reason withheld Jan 26 17:01:14 crc kubenswrapper[4856]: [+]process-running ok Jan 26 17:01:14 crc kubenswrapper[4856]: healthz check failed Jan 26 17:01:14 crc kubenswrapper[4856]: I0126 17:01:14.162212 4856 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9b2g" podUID="85f05bd5-ff83-4d29-9531-ab3499088095" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 17:01:14 crc kubenswrapper[4856]: I0126 17:01:14.232300 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:14 crc kubenswrapper[4856]: E0126 17:01:14.232817 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:14.732793412 +0000 UTC m=+170.686047393 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:14 crc kubenswrapper[4856]: I0126 17:01:14.265059 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-vfm8t" event={"ID":"c8657575-cd22-4ebc-ae9d-4174366985d3","Type":"ContainerStarted","Data":"d0091273fe8514a269a2651f8f1656ec323dcdb4398bc05ab2b74d331928cdf7"} Jan 26 17:01:14 crc kubenswrapper[4856]: I0126 17:01:14.335148 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:14 crc kubenswrapper[4856]: E0126 17:01:14.337495 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:14.837474759 +0000 UTC m=+170.790728730 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:14 crc kubenswrapper[4856]: I0126 17:01:14.461205 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:14 crc kubenswrapper[4856]: E0126 17:01:14.461828 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:14.961803856 +0000 UTC m=+170.915057837 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:14 crc kubenswrapper[4856]: I0126 17:01:14.480753 4856 patch_prober.go:28] interesting pod/downloads-7954f5f757-7l927 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Jan 26 17:01:14 crc kubenswrapper[4856]: I0126 17:01:14.480839 4856 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-7l927" podUID="94291fa4-24a5-499e-8143-89c8784d9284" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Jan 26 17:01:14 crc kubenswrapper[4856]: I0126 17:01:14.482484 4856 patch_prober.go:28] interesting pod/downloads-7954f5f757-7l927 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Jan 26 17:01:14 crc kubenswrapper[4856]: I0126 17:01:14.482570 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-7l927" podUID="94291fa4-24a5-499e-8143-89c8784d9284" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Jan 26 17:01:14 crc kubenswrapper[4856]: I0126 17:01:14.482814 4856 patch_prober.go:28] interesting pod/console-operator-58897d9998-4pbj2 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/readyz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Jan 26 17:01:14 crc kubenswrapper[4856]: I0126 17:01:14.482844 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-4pbj2" podUID="831dc87e-8e14-43d3-a36e-dc7679041ae5" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/readyz\": dial tcp 10.217.0.13:8443: connect: connection refused" Jan 26 17:01:14 crc kubenswrapper[4856]: I0126 17:01:14.482850 4856 patch_prober.go:28] interesting pod/console-operator-58897d9998-4pbj2 container/console-operator namespace/openshift-console-operator: Liveness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Jan 26 17:01:14 crc kubenswrapper[4856]: I0126 17:01:14.482915 4856 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console-operator/console-operator-58897d9998-4pbj2" podUID="831dc87e-8e14-43d3-a36e-dc7679041ae5" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" Jan 26 17:01:14 crc kubenswrapper[4856]: I0126 17:01:14.490196 4856 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-wvttb container/marketplace-operator namespace/openshift-marketplace: Liveness probe status=failure output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" start-of-body= Jan 26 17:01:14 crc kubenswrapper[4856]: I0126 17:01:14.490263 4856 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/marketplace-operator-79b997595-wvttb" podUID="2d37efbf-d18f-486b-9b43-bc4d181af4ca" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" Jan 26 17:01:14 crc kubenswrapper[4856]: I0126 17:01:14.490349 4856 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-wvttb container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" start-of-body= Jan 26 17:01:14 crc kubenswrapper[4856]: I0126 17:01:14.490366 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-wvttb" podUID="2d37efbf-d18f-486b-9b43-bc4d181af4ca" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" Jan 26 17:01:14 crc kubenswrapper[4856]: I0126 17:01:14.494381 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-k662z" Jan 26 17:01:14 crc kubenswrapper[4856]: I0126 17:01:14.564602 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:14 crc kubenswrapper[4856]: E0126 17:01:14.567989 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:15.067966646 +0000 UTC m=+171.021220627 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:14 crc kubenswrapper[4856]: I0126 17:01:14.755050 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-mr7cp" Jan 26 17:01:14 crc kubenswrapper[4856]: I0126 17:01:14.755216 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:14 crc kubenswrapper[4856]: E0126 17:01:14.755946 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:15.255929099 +0000 UTC m=+171.209183090 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:14 crc kubenswrapper[4856]: I0126 17:01:14.759502 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-lndnt" Jan 26 17:01:14 crc kubenswrapper[4856]: I0126 17:01:14.854205 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-nn46h" Jan 26 17:01:14 crc kubenswrapper[4856]: I0126 17:01:14.856811 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:14 crc kubenswrapper[4856]: E0126 17:01:14.858302 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:15.358283297 +0000 UTC m=+171.311537368 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:15 crc kubenswrapper[4856]: I0126 17:01:15.031044 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:15 crc kubenswrapper[4856]: E0126 17:01:15.031439 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:15.531410733 +0000 UTC m=+171.484664714 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:15 crc kubenswrapper[4856]: I0126 17:01:15.057244 4856 patch_prober.go:28] interesting pod/router-default-5444994796-h9b2g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 17:01:15 crc kubenswrapper[4856]: [-]has-synced failed: reason withheld Jan 26 17:01:15 crc kubenswrapper[4856]: [+]process-running ok Jan 26 17:01:15 crc kubenswrapper[4856]: healthz check failed Jan 26 17:01:15 crc kubenswrapper[4856]: I0126 17:01:15.057298 4856 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9b2g" podUID="85f05bd5-ff83-4d29-9531-ab3499088095" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 17:01:15 crc kubenswrapper[4856]: I0126 17:01:15.066229 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 26 17:01:15 crc kubenswrapper[4856]: I0126 17:01:15.194028 4856 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-fpqvc container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 26 17:01:15 crc kubenswrapper[4856]: I0126 17:01:15.194111 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpqvc" podUID="5fe6baed-ab97-4d8a-8be2-6f00f9698136" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 26 17:01:15 crc kubenswrapper[4856]: I0126 17:01:15.194230 4856 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-cb8nk container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.11:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 26 17:01:15 crc kubenswrapper[4856]: I0126 17:01:15.194299 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" podUID="69008ed1-f3e5-400d-852f-adbcd94199f6" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.11:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 26 17:01:15 crc kubenswrapper[4856]: I0126 17:01:15.194678 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:15 crc kubenswrapper[4856]: E0126 17:01:15.195036 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:15.695020428 +0000 UTC m=+171.648274409 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:15 crc kubenswrapper[4856]: I0126 17:01:15.425592 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:15 crc kubenswrapper[4856]: E0126 17:01:15.426604 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:15.926580766 +0000 UTC m=+171.879834747 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:15 crc kubenswrapper[4856]: I0126 17:01:15.443692 4856 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-5bjl7 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Jan 26 17:01:15 crc kubenswrapper[4856]: I0126 17:01:15.443767 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bjl7" podUID="2ba3cf6a-a6be-4108-a155-c8bb530aa037" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Jan 26 17:01:15 crc kubenswrapper[4856]: I0126 17:01:15.527558 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:15 crc kubenswrapper[4856]: E0126 17:01:15.528463 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:16.02844791 +0000 UTC m=+171.981701891 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:15 crc kubenswrapper[4856]: I0126 17:01:15.677839 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:15 crc kubenswrapper[4856]: E0126 17:01:15.678444 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:16.178411173 +0000 UTC m=+172.131665154 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:15 crc kubenswrapper[4856]: I0126 17:01:15.688970 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:15 crc kubenswrapper[4856]: E0126 17:01:15.689985 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:16.189949763 +0000 UTC m=+172.143203744 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:15 crc kubenswrapper[4856]: I0126 17:01:15.795281 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:15 crc kubenswrapper[4856]: E0126 17:01:15.796572 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:16.296516226 +0000 UTC m=+172.249770217 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:15 crc kubenswrapper[4856]: I0126 17:01:15.982000 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:15 crc kubenswrapper[4856]: E0126 17:01:15.982435 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:16.482418728 +0000 UTC m=+172.435672709 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:16 crc kubenswrapper[4856]: I0126 17:01:16.008844 4856 csr.go:261] certificate signing request csr-sr8zt is approved, waiting to be issued Jan 26 17:01:16 crc kubenswrapper[4856]: I0126 17:01:16.008873 4856 csr.go:257] certificate signing request csr-sr8zt is issued Jan 26 17:01:16 crc kubenswrapper[4856]: I0126 17:01:16.390089 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:16 crc kubenswrapper[4856]: E0126 17:01:16.390772 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:16.890738479 +0000 UTC m=+172.843992460 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:16 crc kubenswrapper[4856]: I0126 17:01:16.408594 4856 patch_prober.go:28] interesting pod/router-default-5444994796-h9b2g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 17:01:16 crc kubenswrapper[4856]: [-]has-synced failed: reason withheld Jan 26 17:01:16 crc kubenswrapper[4856]: [+]process-running ok Jan 26 17:01:16 crc kubenswrapper[4856]: healthz check failed Jan 26 17:01:16 crc kubenswrapper[4856]: I0126 17:01:16.408669 4856 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9b2g" podUID="85f05bd5-ff83-4d29-9531-ab3499088095" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 17:01:16 crc kubenswrapper[4856]: I0126 17:01:16.527309 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:16 crc kubenswrapper[4856]: E0126 17:01:16.527918 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:17.027890633 +0000 UTC m=+172.981144614 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:16 crc kubenswrapper[4856]: I0126 17:01:16.705905 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:16 crc kubenswrapper[4856]: E0126 17:01:16.706823 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:17.206804389 +0000 UTC m=+173.160058360 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:16 crc kubenswrapper[4856]: I0126 17:01:16.750025 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-vfm8t" event={"ID":"c8657575-cd22-4ebc-ae9d-4174366985d3","Type":"ContainerStarted","Data":"aaf3beaf7acc57a58fac4f6add0ba1742c11399c8e5ad40f941c85d9c4ee334e"} Jan 26 17:01:16 crc kubenswrapper[4856]: I0126 17:01:16.776887 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"d8c22047-144c-402a-80c5-c206539b6826","Type":"ContainerStarted","Data":"eb80e342e945ab720a056481e6d379636786d87ebb81ecfa7bcd84ffb36388ff"} Jan 26 17:01:16 crc kubenswrapper[4856]: I0126 17:01:16.817395 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:16 crc kubenswrapper[4856]: E0126 17:01:16.817798 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:17.317783702 +0000 UTC m=+173.271037683 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:16 crc kubenswrapper[4856]: I0126 17:01:16.863770 4856 generic.go:334] "Generic (PLEG): container finished" podID="56258114-1bee-4516-ab71-f60d15a9635d" containerID="688d67199a7c309449abcb4f65cea024a314835283d336314d00644e67f44daf" exitCode=0 Jan 26 17:01:16 crc kubenswrapper[4856]: I0126 17:01:16.863848 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"56258114-1bee-4516-ab71-f60d15a9635d","Type":"ContainerDied","Data":"688d67199a7c309449abcb4f65cea024a314835283d336314d00644e67f44daf"} Jan 26 17:01:16 crc kubenswrapper[4856]: I0126 17:01:16.920044 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:16 crc kubenswrapper[4856]: E0126 17:01:16.920364 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:17.420345337 +0000 UTC m=+173.373599318 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:16 crc kubenswrapper[4856]: I0126 17:01:16.997468 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=14.997446829 podStartE2EDuration="14.997446829s" podCreationTimestamp="2026-01-26 17:01:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:01:16.943435307 +0000 UTC m=+172.896689288" watchObservedRunningTime="2026-01-26 17:01:16.997446829 +0000 UTC m=+172.950700810" Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.009693 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-01-26 16:56:15 +0000 UTC, rotation deadline is 2026-10-11 11:23:10.578419877 +0000 UTC Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.009755 4856 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6186h21m53.568668175s for next certificate rotation Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.025201 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:17 crc kubenswrapper[4856]: E0126 17:01:17.025567 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:17.525555108 +0000 UTC m=+173.478809079 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.152022 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:17 crc kubenswrapper[4856]: E0126 17:01:17.152269 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:17.652237034 +0000 UTC m=+173.605491015 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.152324 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:17 crc kubenswrapper[4856]: E0126 17:01:17.152725 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:17.652715968 +0000 UTC m=+173.605969949 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.244082 4856 patch_prober.go:28] interesting pod/router-default-5444994796-h9b2g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 17:01:17 crc kubenswrapper[4856]: [-]has-synced failed: reason withheld Jan 26 17:01:17 crc kubenswrapper[4856]: [+]process-running ok Jan 26 17:01:17 crc kubenswrapper[4856]: healthz check failed Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.244468 4856 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9b2g" podUID="85f05bd5-ff83-4d29-9531-ab3499088095" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.253621 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:17 crc kubenswrapper[4856]: E0126 17:01:17.253771 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:17.753748148 +0000 UTC m=+173.707002129 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.253854 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:17 crc kubenswrapper[4856]: E0126 17:01:17.254217 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:17.754207871 +0000 UTC m=+173.707461922 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.264181 4856 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6cghs" Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.292780 4856 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-6cghs container/oauth-apiserver namespace/openshift-oauth-apiserver: Readiness probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Jan 26 17:01:17 crc kubenswrapper[4856]: [+]log ok Jan 26 17:01:17 crc kubenswrapper[4856]: [+]etcd ok Jan 26 17:01:17 crc kubenswrapper[4856]: [+]etcd-readiness ok Jan 26 17:01:17 crc kubenswrapper[4856]: [+]poststarthook/start-apiserver-admission-initializer ok Jan 26 17:01:17 crc kubenswrapper[4856]: [-]informer-sync failed: reason withheld Jan 26 17:01:17 crc kubenswrapper[4856]: [+]poststarthook/generic-apiserver-start-informers ok Jan 26 17:01:17 crc kubenswrapper[4856]: [+]poststarthook/max-in-flight-filter ok Jan 26 17:01:17 crc kubenswrapper[4856]: [+]poststarthook/storage-object-count-tracker-hook ok Jan 26 17:01:17 crc kubenswrapper[4856]: [+]poststarthook/openshift.io-StartUserInformer ok Jan 26 17:01:17 crc kubenswrapper[4856]: [+]poststarthook/openshift.io-StartOAuthInformer ok Jan 26 17:01:17 crc kubenswrapper[4856]: [+]poststarthook/openshift.io-StartTokenTimeoutUpdater ok Jan 26 17:01:17 crc kubenswrapper[4856]: [+]shutdown ok Jan 26 17:01:17 crc kubenswrapper[4856]: readyz check failed Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.292844 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6cghs" podUID="81c2f96b-55e0-483b-b72c-df7e156e9218" containerName="oauth-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.344348 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5bjl7" Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.354705 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:17 crc kubenswrapper[4856]: E0126 17:01:17.354857 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:17.854834899 +0000 UTC m=+173.808088900 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.355026 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:17 crc kubenswrapper[4856]: E0126 17:01:17.356065 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:17.856043164 +0000 UTC m=+173.809297235 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.455762 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:17 crc kubenswrapper[4856]: E0126 17:01:17.455909 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:17.955886389 +0000 UTC m=+173.909140370 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.456005 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:17 crc kubenswrapper[4856]: E0126 17:01:17.456289 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:17.95627707 +0000 UTC m=+173.909531051 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.498337 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-txmdl"] Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.499764 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-txmdl" Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.509232 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.529285 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-txmdl"] Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.557193 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:17 crc kubenswrapper[4856]: E0126 17:01:17.557305 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:18.057286409 +0000 UTC m=+174.010540380 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.557602 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40a27476-22b1-4083-990e-66e70ccdaf4c-utilities\") pod \"certified-operators-txmdl\" (UID: \"40a27476-22b1-4083-990e-66e70ccdaf4c\") " pod="openshift-marketplace/certified-operators-txmdl" Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.557788 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tf49w\" (UniqueName: \"kubernetes.io/projected/40a27476-22b1-4083-990e-66e70ccdaf4c-kube-api-access-tf49w\") pod \"certified-operators-txmdl\" (UID: \"40a27476-22b1-4083-990e-66e70ccdaf4c\") " pod="openshift-marketplace/certified-operators-txmdl" Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.557872 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40a27476-22b1-4083-990e-66e70ccdaf4c-catalog-content\") pod \"certified-operators-txmdl\" (UID: \"40a27476-22b1-4083-990e-66e70ccdaf4c\") " pod="openshift-marketplace/certified-operators-txmdl" Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.557962 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:17 crc kubenswrapper[4856]: E0126 17:01:17.558371 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:18.05835126 +0000 UTC m=+174.011605241 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.659188 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.659418 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40a27476-22b1-4083-990e-66e70ccdaf4c-utilities\") pod \"certified-operators-txmdl\" (UID: \"40a27476-22b1-4083-990e-66e70ccdaf4c\") " pod="openshift-marketplace/certified-operators-txmdl" Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.659472 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tf49w\" (UniqueName: \"kubernetes.io/projected/40a27476-22b1-4083-990e-66e70ccdaf4c-kube-api-access-tf49w\") pod \"certified-operators-txmdl\" (UID: \"40a27476-22b1-4083-990e-66e70ccdaf4c\") " pod="openshift-marketplace/certified-operators-txmdl" Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.659505 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40a27476-22b1-4083-990e-66e70ccdaf4c-catalog-content\") pod \"certified-operators-txmdl\" (UID: \"40a27476-22b1-4083-990e-66e70ccdaf4c\") " pod="openshift-marketplace/certified-operators-txmdl" Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.660015 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40a27476-22b1-4083-990e-66e70ccdaf4c-catalog-content\") pod \"certified-operators-txmdl\" (UID: \"40a27476-22b1-4083-990e-66e70ccdaf4c\") " pod="openshift-marketplace/certified-operators-txmdl" Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.660213 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40a27476-22b1-4083-990e-66e70ccdaf4c-utilities\") pod \"certified-operators-txmdl\" (UID: \"40a27476-22b1-4083-990e-66e70ccdaf4c\") " pod="openshift-marketplace/certified-operators-txmdl" Jan 26 17:01:17 crc kubenswrapper[4856]: E0126 17:01:17.660391 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:18.160369809 +0000 UTC m=+174.113623790 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.661200 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-n8hp2"] Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.663921 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n8hp2" Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.674959 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.700513 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n8hp2"] Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.708931 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tf49w\" (UniqueName: \"kubernetes.io/projected/40a27476-22b1-4083-990e-66e70ccdaf4c-kube-api-access-tf49w\") pod \"certified-operators-txmdl\" (UID: \"40a27476-22b1-4083-990e-66e70ccdaf4c\") " pod="openshift-marketplace/certified-operators-txmdl" Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.760661 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6086d4b-faeb-4a12-8e6a-2a178dfe374c-utilities\") pod \"community-operators-n8hp2\" (UID: \"a6086d4b-faeb-4a12-8e6a-2a178dfe374c\") " pod="openshift-marketplace/community-operators-n8hp2" Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.760792 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7mvx\" (UniqueName: \"kubernetes.io/projected/a6086d4b-faeb-4a12-8e6a-2a178dfe374c-kube-api-access-x7mvx\") pod \"community-operators-n8hp2\" (UID: \"a6086d4b-faeb-4a12-8e6a-2a178dfe374c\") " pod="openshift-marketplace/community-operators-n8hp2" Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.760833 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.760862 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6086d4b-faeb-4a12-8e6a-2a178dfe374c-catalog-content\") pod \"community-operators-n8hp2\" (UID: \"a6086d4b-faeb-4a12-8e6a-2a178dfe374c\") " pod="openshift-marketplace/community-operators-n8hp2" Jan 26 17:01:17 crc kubenswrapper[4856]: E0126 17:01:17.761232 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:18.261216933 +0000 UTC m=+174.214470914 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.815496 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-txmdl" Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.823579 4856 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.847181 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-62nhd"] Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.848170 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-62nhd" Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.861424 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.861576 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6086d4b-faeb-4a12-8e6a-2a178dfe374c-catalog-content\") pod \"community-operators-n8hp2\" (UID: \"a6086d4b-faeb-4a12-8e6a-2a178dfe374c\") " pod="openshift-marketplace/community-operators-n8hp2" Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.861618 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6086d4b-faeb-4a12-8e6a-2a178dfe374c-utilities\") pod \"community-operators-n8hp2\" (UID: \"a6086d4b-faeb-4a12-8e6a-2a178dfe374c\") " pod="openshift-marketplace/community-operators-n8hp2" Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.861661 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s92lp\" (UniqueName: \"kubernetes.io/projected/7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766-kube-api-access-s92lp\") pod \"certified-operators-62nhd\" (UID: \"7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766\") " pod="openshift-marketplace/certified-operators-62nhd" Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.861679 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766-catalog-content\") pod \"certified-operators-62nhd\" (UID: \"7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766\") " pod="openshift-marketplace/certified-operators-62nhd" Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.861709 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766-utilities\") pod \"certified-operators-62nhd\" (UID: \"7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766\") " pod="openshift-marketplace/certified-operators-62nhd" Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.861732 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7mvx\" (UniqueName: \"kubernetes.io/projected/a6086d4b-faeb-4a12-8e6a-2a178dfe374c-kube-api-access-x7mvx\") pod \"community-operators-n8hp2\" (UID: \"a6086d4b-faeb-4a12-8e6a-2a178dfe374c\") " pod="openshift-marketplace/community-operators-n8hp2" Jan 26 17:01:17 crc kubenswrapper[4856]: E0126 17:01:17.862144 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:18.362125188 +0000 UTC m=+174.315379169 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.862165 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6086d4b-faeb-4a12-8e6a-2a178dfe374c-utilities\") pod \"community-operators-n8hp2\" (UID: \"a6086d4b-faeb-4a12-8e6a-2a178dfe374c\") " pod="openshift-marketplace/community-operators-n8hp2" Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.862502 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6086d4b-faeb-4a12-8e6a-2a178dfe374c-catalog-content\") pod \"community-operators-n8hp2\" (UID: \"a6086d4b-faeb-4a12-8e6a-2a178dfe374c\") " pod="openshift-marketplace/community-operators-n8hp2" Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.878981 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"d8c22047-144c-402a-80c5-c206539b6826","Type":"ContainerStarted","Data":"1ebaf120a54fd03aee32564585380d095168008ce25b9456bb3082fadde275b6"} Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.881484 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-62nhd"] Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.932743 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=11.93272413 podStartE2EDuration="11.93272413s" podCreationTimestamp="2026-01-26 17:01:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:01:17.929892207 +0000 UTC m=+173.883146198" watchObservedRunningTime="2026-01-26 17:01:17.93272413 +0000 UTC m=+173.885978111" Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.939614 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7mvx\" (UniqueName: \"kubernetes.io/projected/a6086d4b-faeb-4a12-8e6a-2a178dfe374c-kube-api-access-x7mvx\") pod \"community-operators-n8hp2\" (UID: \"a6086d4b-faeb-4a12-8e6a-2a178dfe374c\") " pod="openshift-marketplace/community-operators-n8hp2" Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.950629 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-vfm8t" event={"ID":"c8657575-cd22-4ebc-ae9d-4174366985d3","Type":"ContainerStarted","Data":"3af0c7fd8b578466852ece6c83fad0a731244f1ccf3b85c06b6618e8ba6a4cad"} Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.965291 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s92lp\" (UniqueName: \"kubernetes.io/projected/7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766-kube-api-access-s92lp\") pod \"certified-operators-62nhd\" (UID: \"7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766\") " pod="openshift-marketplace/certified-operators-62nhd" Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.965337 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766-catalog-content\") pod \"certified-operators-62nhd\" (UID: \"7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766\") " pod="openshift-marketplace/certified-operators-62nhd" Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.965377 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766-utilities\") pod \"certified-operators-62nhd\" (UID: \"7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766\") " pod="openshift-marketplace/certified-operators-62nhd" Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.965415 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.966101 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766-catalog-content\") pod \"certified-operators-62nhd\" (UID: \"7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766\") " pod="openshift-marketplace/certified-operators-62nhd" Jan 26 17:01:17 crc kubenswrapper[4856]: E0126 17:01:17.966225 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:18.466212038 +0000 UTC m=+174.419466019 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.966552 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766-utilities\") pod \"certified-operators-62nhd\" (UID: \"7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766\") " pod="openshift-marketplace/certified-operators-62nhd" Jan 26 17:01:17 crc kubenswrapper[4856]: I0126 17:01:17.983519 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n8hp2" Jan 26 17:01:18 crc kubenswrapper[4856]: I0126 17:01:18.011294 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s92lp\" (UniqueName: \"kubernetes.io/projected/7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766-kube-api-access-s92lp\") pod \"certified-operators-62nhd\" (UID: \"7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766\") " pod="openshift-marketplace/certified-operators-62nhd" Jan 26 17:01:18 crc kubenswrapper[4856]: I0126 17:01:18.060847 4856 patch_prober.go:28] interesting pod/router-default-5444994796-h9b2g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 17:01:18 crc kubenswrapper[4856]: [-]has-synced failed: reason withheld Jan 26 17:01:18 crc kubenswrapper[4856]: [+]process-running ok Jan 26 17:01:18 crc kubenswrapper[4856]: healthz check failed Jan 26 17:01:18 crc kubenswrapper[4856]: I0126 17:01:18.060920 4856 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9b2g" podUID="85f05bd5-ff83-4d29-9531-ab3499088095" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 17:01:18 crc kubenswrapper[4856]: I0126 17:01:18.073135 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:18 crc kubenswrapper[4856]: E0126 17:01:18.073674 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:18.573640556 +0000 UTC m=+174.526894537 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:18 crc kubenswrapper[4856]: I0126 17:01:18.080262 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-vfm8t" podStartSLOduration=27.08024018 podStartE2EDuration="27.08024018s" podCreationTimestamp="2026-01-26 17:00:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:01:18.009902766 +0000 UTC m=+173.963156757" watchObservedRunningTime="2026-01-26 17:01:18.08024018 +0000 UTC m=+174.033494161" Jan 26 17:01:18 crc kubenswrapper[4856]: I0126 17:01:18.080470 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-qgjjd"] Jan 26 17:01:18 crc kubenswrapper[4856]: I0126 17:01:18.081720 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qgjjd" Jan 26 17:01:18 crc kubenswrapper[4856]: I0126 17:01:18.115201 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qgjjd"] Jan 26 17:01:18 crc kubenswrapper[4856]: I0126 17:01:18.175240 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89cf05de-642b-4574-9f79-45e7a3d4afa3-utilities\") pod \"community-operators-qgjjd\" (UID: \"89cf05de-642b-4574-9f79-45e7a3d4afa3\") " pod="openshift-marketplace/community-operators-qgjjd" Jan 26 17:01:18 crc kubenswrapper[4856]: I0126 17:01:18.175296 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:18 crc kubenswrapper[4856]: I0126 17:01:18.175361 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89cf05de-642b-4574-9f79-45e7a3d4afa3-catalog-content\") pod \"community-operators-qgjjd\" (UID: \"89cf05de-642b-4574-9f79-45e7a3d4afa3\") " pod="openshift-marketplace/community-operators-qgjjd" Jan 26 17:01:18 crc kubenswrapper[4856]: I0126 17:01:18.175419 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtxjz\" (UniqueName: \"kubernetes.io/projected/89cf05de-642b-4574-9f79-45e7a3d4afa3-kube-api-access-gtxjz\") pod \"community-operators-qgjjd\" (UID: \"89cf05de-642b-4574-9f79-45e7a3d4afa3\") " pod="openshift-marketplace/community-operators-qgjjd" Jan 26 17:01:18 crc kubenswrapper[4856]: E0126 17:01:18.175748 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:18.675726496 +0000 UTC m=+174.628980537 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:18 crc kubenswrapper[4856]: I0126 17:01:18.191851 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-62nhd" Jan 26 17:01:18 crc kubenswrapper[4856]: I0126 17:01:18.280184 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:18 crc kubenswrapper[4856]: I0126 17:01:18.280389 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89cf05de-642b-4574-9f79-45e7a3d4afa3-catalog-content\") pod \"community-operators-qgjjd\" (UID: \"89cf05de-642b-4574-9f79-45e7a3d4afa3\") " pod="openshift-marketplace/community-operators-qgjjd" Jan 26 17:01:18 crc kubenswrapper[4856]: I0126 17:01:18.280438 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtxjz\" (UniqueName: \"kubernetes.io/projected/89cf05de-642b-4574-9f79-45e7a3d4afa3-kube-api-access-gtxjz\") pod \"community-operators-qgjjd\" (UID: \"89cf05de-642b-4574-9f79-45e7a3d4afa3\") " pod="openshift-marketplace/community-operators-qgjjd" Jan 26 17:01:18 crc kubenswrapper[4856]: I0126 17:01:18.280472 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89cf05de-642b-4574-9f79-45e7a3d4afa3-utilities\") pod \"community-operators-qgjjd\" (UID: \"89cf05de-642b-4574-9f79-45e7a3d4afa3\") " pod="openshift-marketplace/community-operators-qgjjd" Jan 26 17:01:18 crc kubenswrapper[4856]: I0126 17:01:18.280991 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89cf05de-642b-4574-9f79-45e7a3d4afa3-utilities\") pod \"community-operators-qgjjd\" (UID: \"89cf05de-642b-4574-9f79-45e7a3d4afa3\") " pod="openshift-marketplace/community-operators-qgjjd" Jan 26 17:01:18 crc kubenswrapper[4856]: E0126 17:01:18.281365 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:18.781348741 +0000 UTC m=+174.734602722 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:18 crc kubenswrapper[4856]: I0126 17:01:18.283162 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89cf05de-642b-4574-9f79-45e7a3d4afa3-catalog-content\") pod \"community-operators-qgjjd\" (UID: \"89cf05de-642b-4574-9f79-45e7a3d4afa3\") " pod="openshift-marketplace/community-operators-qgjjd" Jan 26 17:01:18 crc kubenswrapper[4856]: I0126 17:01:18.334699 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtxjz\" (UniqueName: \"kubernetes.io/projected/89cf05de-642b-4574-9f79-45e7a3d4afa3-kube-api-access-gtxjz\") pod \"community-operators-qgjjd\" (UID: \"89cf05de-642b-4574-9f79-45e7a3d4afa3\") " pod="openshift-marketplace/community-operators-qgjjd" Jan 26 17:01:18 crc kubenswrapper[4856]: I0126 17:01:18.381323 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12e50462-28e6-4531-ada4-e652310e6cce-metrics-certs\") pod \"network-metrics-daemon-295wr\" (UID: \"12e50462-28e6-4531-ada4-e652310e6cce\") " pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 17:01:18 crc kubenswrapper[4856]: I0126 17:01:18.381430 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:18 crc kubenswrapper[4856]: E0126 17:01:18.381748 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:18.881735661 +0000 UTC m=+174.834989632 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:18 crc kubenswrapper[4856]: I0126 17:01:18.388148 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/12e50462-28e6-4531-ada4-e652310e6cce-metrics-certs\") pod \"network-metrics-daemon-295wr\" (UID: \"12e50462-28e6-4531-ada4-e652310e6cce\") " pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 17:01:18 crc kubenswrapper[4856]: I0126 17:01:18.407126 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qgjjd" Jan 26 17:01:18 crc kubenswrapper[4856]: I0126 17:01:18.485225 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:18 crc kubenswrapper[4856]: E0126 17:01:18.485566 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:18.985547033 +0000 UTC m=+174.938801014 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:18 crc kubenswrapper[4856]: I0126 17:01:18.540635 4856 patch_prober.go:28] interesting pod/apiserver-76f77b778f-6rlxp container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Jan 26 17:01:18 crc kubenswrapper[4856]: [+]log ok Jan 26 17:01:18 crc kubenswrapper[4856]: [+]etcd ok Jan 26 17:01:18 crc kubenswrapper[4856]: [+]poststarthook/start-apiserver-admission-initializer ok Jan 26 17:01:18 crc kubenswrapper[4856]: [+]poststarthook/generic-apiserver-start-informers ok Jan 26 17:01:18 crc kubenswrapper[4856]: [+]poststarthook/max-in-flight-filter ok Jan 26 17:01:18 crc kubenswrapper[4856]: [+]poststarthook/storage-object-count-tracker-hook ok Jan 26 17:01:18 crc kubenswrapper[4856]: [+]poststarthook/image.openshift.io-apiserver-caches ok Jan 26 17:01:18 crc kubenswrapper[4856]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Jan 26 17:01:18 crc kubenswrapper[4856]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Jan 26 17:01:18 crc kubenswrapper[4856]: [+]poststarthook/project.openshift.io-projectcache ok Jan 26 17:01:18 crc kubenswrapper[4856]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Jan 26 17:01:18 crc kubenswrapper[4856]: [+]poststarthook/openshift.io-startinformers ok Jan 26 17:01:18 crc kubenswrapper[4856]: [+]poststarthook/openshift.io-restmapperupdater ok Jan 26 17:01:18 crc kubenswrapper[4856]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Jan 26 17:01:18 crc kubenswrapper[4856]: livez check failed Jan 26 17:01:18 crc kubenswrapper[4856]: I0126 17:01:18.540706 4856 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" podUID="a6d331bd-2db3-4319-9f5c-db56d408d9e3" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 17:01:18 crc kubenswrapper[4856]: I0126 17:01:18.542837 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-295wr" Jan 26 17:01:18 crc kubenswrapper[4856]: I0126 17:01:18.586505 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:18 crc kubenswrapper[4856]: E0126 17:01:18.586966 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:19.086950213 +0000 UTC m=+175.040204194 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:18 crc kubenswrapper[4856]: I0126 17:01:18.690654 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:18 crc kubenswrapper[4856]: E0126 17:01:18.690956 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:19.19093661 +0000 UTC m=+175.144190591 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:18 crc kubenswrapper[4856]: I0126 17:01:18.703639 4856 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-01-26T17:01:17.82385409Z","Handler":null,"Name":""} Jan 26 17:01:18 crc kubenswrapper[4856]: I0126 17:01:18.755830 4856 patch_prober.go:28] interesting pod/apiserver-76f77b778f-6rlxp container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Jan 26 17:01:18 crc kubenswrapper[4856]: [+]log ok Jan 26 17:01:18 crc kubenswrapper[4856]: [+]etcd ok Jan 26 17:01:18 crc kubenswrapper[4856]: [+]poststarthook/start-apiserver-admission-initializer ok Jan 26 17:01:18 crc kubenswrapper[4856]: [+]poststarthook/generic-apiserver-start-informers ok Jan 26 17:01:18 crc kubenswrapper[4856]: [+]poststarthook/max-in-flight-filter ok Jan 26 17:01:18 crc kubenswrapper[4856]: [+]poststarthook/storage-object-count-tracker-hook ok Jan 26 17:01:18 crc kubenswrapper[4856]: [+]poststarthook/image.openshift.io-apiserver-caches ok Jan 26 17:01:18 crc kubenswrapper[4856]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Jan 26 17:01:18 crc kubenswrapper[4856]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Jan 26 17:01:18 crc kubenswrapper[4856]: [+]poststarthook/project.openshift.io-projectcache ok Jan 26 17:01:18 crc kubenswrapper[4856]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Jan 26 17:01:18 crc kubenswrapper[4856]: [+]poststarthook/openshift.io-startinformers ok Jan 26 17:01:18 crc kubenswrapper[4856]: [+]poststarthook/openshift.io-restmapperupdater ok Jan 26 17:01:18 crc kubenswrapper[4856]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Jan 26 17:01:18 crc kubenswrapper[4856]: livez check failed Jan 26 17:01:18 crc kubenswrapper[4856]: I0126 17:01:18.756148 4856 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" podUID="a6d331bd-2db3-4319-9f5c-db56d408d9e3" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 17:01:18 crc kubenswrapper[4856]: I0126 17:01:18.792315 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:18 crc kubenswrapper[4856]: E0126 17:01:18.792666 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 17:01:19.292632239 +0000 UTC m=+175.245886220 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-wxbdh" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:18 crc kubenswrapper[4856]: I0126 17:01:18.893114 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:18 crc kubenswrapper[4856]: E0126 17:01:18.893762 4856 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 17:01:19.39373925 +0000 UTC m=+175.346993231 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 17:01:18 crc kubenswrapper[4856]: I0126 17:01:18.911819 4856 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Jan 26 17:01:18 crc kubenswrapper[4856]: I0126 17:01:18.911874 4856 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Jan 26 17:01:18 crc kubenswrapper[4856]: I0126 17:01:18.972672 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 26 17:01:18 crc kubenswrapper[4856]: I0126 17:01:18.985614 4856 generic.go:334] "Generic (PLEG): container finished" podID="d8c22047-144c-402a-80c5-c206539b6826" containerID="1ebaf120a54fd03aee32564585380d095168008ce25b9456bb3082fadde275b6" exitCode=0 Jan 26 17:01:18 crc kubenswrapper[4856]: I0126 17:01:18.985764 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"d8c22047-144c-402a-80c5-c206539b6826","Type":"ContainerDied","Data":"1ebaf120a54fd03aee32564585380d095168008ce25b9456bb3082fadde275b6"} Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.008212 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/56258114-1bee-4516-ab71-f60d15a9635d-kube-api-access\") pod \"56258114-1bee-4516-ab71-f60d15a9635d\" (UID: \"56258114-1bee-4516-ab71-f60d15a9635d\") " Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.008399 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/56258114-1bee-4516-ab71-f60d15a9635d-kubelet-dir\") pod \"56258114-1bee-4516-ab71-f60d15a9635d\" (UID: \"56258114-1bee-4516-ab71-f60d15a9635d\") " Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.008587 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.010313 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/56258114-1bee-4516-ab71-f60d15a9635d-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "56258114-1bee-4516-ab71-f60d15a9635d" (UID: "56258114-1bee-4516-ab71-f60d15a9635d"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.024596 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.024599 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"56258114-1bee-4516-ab71-f60d15a9635d","Type":"ContainerDied","Data":"02721e310b92647ac526a2662cb5ceecb1c039ddbbb08ec7de0c0bb07775f5b7"} Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.024653 4856 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="02721e310b92647ac526a2662cb5ceecb1c039ddbbb08ec7de0c0bb07775f5b7" Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.030909 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56258114-1bee-4516-ab71-f60d15a9635d-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "56258114-1bee-4516-ab71-f60d15a9635d" (UID: "56258114-1bee-4516-ab71-f60d15a9635d"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.057143 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6cghs" Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.063437 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-txmdl"] Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.136398 4856 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/56258114-1bee-4516-ab71-f60d15a9635d-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.136423 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/56258114-1bee-4516-ab71-f60d15a9635d-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.138629 4856 patch_prober.go:28] interesting pod/router-default-5444994796-h9b2g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 17:01:19 crc kubenswrapper[4856]: [-]has-synced failed: reason withheld Jan 26 17:01:19 crc kubenswrapper[4856]: [+]process-running ok Jan 26 17:01:19 crc kubenswrapper[4856]: healthz check failed Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.138687 4856 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9b2g" podUID="85f05bd5-ff83-4d29-9531-ab3499088095" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.151906 4856 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.151975 4856 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.190998 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n8hp2"] Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.337690 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-62nhd"] Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.340334 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-wxbdh\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.422093 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.456147 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.495735 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4kwt4"] Jan 26 17:01:19 crc kubenswrapper[4856]: E0126 17:01:19.496113 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56258114-1bee-4516-ab71-f60d15a9635d" containerName="pruner" Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.496137 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="56258114-1bee-4516-ab71-f60d15a9635d" containerName="pruner" Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.496274 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="56258114-1bee-4516-ab71-f60d15a9635d" containerName="pruner" Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.497356 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4kwt4" Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.515709 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.521245 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qgjjd"] Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.534304 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4kwt4"] Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.625416 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.630511 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lndnt"] Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.630820 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-lndnt" podUID="1afc0f4c-e02d-4a70-aaba-e761e8c04eee" containerName="controller-manager" containerID="cri-o://e9e54e2a4a2266ca4148b11cb38df08f87c2f2ccd87dc3343d147862786c16e2" gracePeriod=30 Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.650975 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpqvc"] Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.651243 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpqvc" podUID="5fe6baed-ab97-4d8a-8be2-6f00f9698136" containerName="route-controller-manager" containerID="cri-o://56133c3e036efeb9590dc043f9b9af766fce603e2c50cfdca46be37466b88f62" gracePeriod=30 Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.659480 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6944fc9-b8d7-4013-8702-b5765c410a0b-catalog-content\") pod \"redhat-marketplace-4kwt4\" (UID: \"d6944fc9-b8d7-4013-8702-b5765c410a0b\") " pod="openshift-marketplace/redhat-marketplace-4kwt4" Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.659562 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qzgz\" (UniqueName: \"kubernetes.io/projected/d6944fc9-b8d7-4013-8702-b5765c410a0b-kube-api-access-2qzgz\") pod \"redhat-marketplace-4kwt4\" (UID: \"d6944fc9-b8d7-4013-8702-b5765c410a0b\") " pod="openshift-marketplace/redhat-marketplace-4kwt4" Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.659592 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6944fc9-b8d7-4013-8702-b5765c410a0b-utilities\") pod \"redhat-marketplace-4kwt4\" (UID: \"d6944fc9-b8d7-4013-8702-b5765c410a0b\") " pod="openshift-marketplace/redhat-marketplace-4kwt4" Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.674391 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpqvc" Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.764555 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6944fc9-b8d7-4013-8702-b5765c410a0b-catalog-content\") pod \"redhat-marketplace-4kwt4\" (UID: \"d6944fc9-b8d7-4013-8702-b5765c410a0b\") " pod="openshift-marketplace/redhat-marketplace-4kwt4" Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.764643 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qzgz\" (UniqueName: \"kubernetes.io/projected/d6944fc9-b8d7-4013-8702-b5765c410a0b-kube-api-access-2qzgz\") pod \"redhat-marketplace-4kwt4\" (UID: \"d6944fc9-b8d7-4013-8702-b5765c410a0b\") " pod="openshift-marketplace/redhat-marketplace-4kwt4" Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.764687 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6944fc9-b8d7-4013-8702-b5765c410a0b-utilities\") pod \"redhat-marketplace-4kwt4\" (UID: \"d6944fc9-b8d7-4013-8702-b5765c410a0b\") " pod="openshift-marketplace/redhat-marketplace-4kwt4" Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.765329 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6944fc9-b8d7-4013-8702-b5765c410a0b-utilities\") pod \"redhat-marketplace-4kwt4\" (UID: \"d6944fc9-b8d7-4013-8702-b5765c410a0b\") " pod="openshift-marketplace/redhat-marketplace-4kwt4" Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.765618 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6944fc9-b8d7-4013-8702-b5765c410a0b-catalog-content\") pod \"redhat-marketplace-4kwt4\" (UID: \"d6944fc9-b8d7-4013-8702-b5765c410a0b\") " pod="openshift-marketplace/redhat-marketplace-4kwt4" Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.791069 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-295wr"] Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.824125 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qzgz\" (UniqueName: \"kubernetes.io/projected/d6944fc9-b8d7-4013-8702-b5765c410a0b-kube-api-access-2qzgz\") pod \"redhat-marketplace-4kwt4\" (UID: \"d6944fc9-b8d7-4013-8702-b5765c410a0b\") " pod="openshift-marketplace/redhat-marketplace-4kwt4" Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.850873 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-g8bgt"] Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.851932 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g8bgt" Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.892913 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4kwt4" Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.942392 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-g8bgt"] Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.967202 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcbvc\" (UniqueName: \"kubernetes.io/projected/0d7eb7b8-63ae-493a-850b-0b9f3b42e927-kube-api-access-dcbvc\") pod \"redhat-marketplace-g8bgt\" (UID: \"0d7eb7b8-63ae-493a-850b-0b9f3b42e927\") " pod="openshift-marketplace/redhat-marketplace-g8bgt" Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.967267 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d7eb7b8-63ae-493a-850b-0b9f3b42e927-utilities\") pod \"redhat-marketplace-g8bgt\" (UID: \"0d7eb7b8-63ae-493a-850b-0b9f3b42e927\") " pod="openshift-marketplace/redhat-marketplace-g8bgt" Jan 26 17:01:19 crc kubenswrapper[4856]: I0126 17:01:19.967324 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d7eb7b8-63ae-493a-850b-0b9f3b42e927-catalog-content\") pod \"redhat-marketplace-g8bgt\" (UID: \"0d7eb7b8-63ae-493a-850b-0b9f3b42e927\") " pod="openshift-marketplace/redhat-marketplace-g8bgt" Jan 26 17:01:20 crc kubenswrapper[4856]: I0126 17:01:20.023226 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-wxbdh"] Jan 26 17:01:20 crc kubenswrapper[4856]: I0126 17:01:20.041663 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n8hp2" event={"ID":"a6086d4b-faeb-4a12-8e6a-2a178dfe374c","Type":"ContainerStarted","Data":"61bc611402534dad5a09a8edd4e25038026dc1769890ea9d2407a69eb9c888af"} Jan 26 17:01:20 crc kubenswrapper[4856]: I0126 17:01:20.045328 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-295wr" event={"ID":"12e50462-28e6-4531-ada4-e652310e6cce","Type":"ContainerStarted","Data":"55ca52361a33fc96ae9820708b118cc821392b787ebe52650ff20c003ce403f1"} Jan 26 17:01:20 crc kubenswrapper[4856]: I0126 17:01:20.054715 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-txmdl" event={"ID":"40a27476-22b1-4083-990e-66e70ccdaf4c","Type":"ContainerStarted","Data":"894929ba59d66c867404dc7094d1e4c1b977bab79b099140b34c889e7b66ae16"} Jan 26 17:01:20 crc kubenswrapper[4856]: I0126 17:01:20.059653 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qgjjd" event={"ID":"89cf05de-642b-4574-9f79-45e7a3d4afa3","Type":"ContainerStarted","Data":"e31b957fac8983059a89e5a7867c6294be7613d1c35b810e6c7face168eea509"} Jan 26 17:01:20 crc kubenswrapper[4856]: I0126 17:01:20.063852 4856 patch_prober.go:28] interesting pod/router-default-5444994796-h9b2g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 17:01:20 crc kubenswrapper[4856]: [-]has-synced failed: reason withheld Jan 26 17:01:20 crc kubenswrapper[4856]: [+]process-running ok Jan 26 17:01:20 crc kubenswrapper[4856]: healthz check failed Jan 26 17:01:20 crc kubenswrapper[4856]: I0126 17:01:20.063890 4856 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9b2g" podUID="85f05bd5-ff83-4d29-9531-ab3499088095" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 17:01:20 crc kubenswrapper[4856]: I0126 17:01:20.064131 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-62nhd" event={"ID":"7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766","Type":"ContainerStarted","Data":"54783f51c7d33737624b9dffb5983a3ed107d31d30f3fd03ab73e5627dfd4bfd"} Jan 26 17:01:20 crc kubenswrapper[4856]: I0126 17:01:20.068924 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d7eb7b8-63ae-493a-850b-0b9f3b42e927-catalog-content\") pod \"redhat-marketplace-g8bgt\" (UID: \"0d7eb7b8-63ae-493a-850b-0b9f3b42e927\") " pod="openshift-marketplace/redhat-marketplace-g8bgt" Jan 26 17:01:20 crc kubenswrapper[4856]: I0126 17:01:20.069032 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcbvc\" (UniqueName: \"kubernetes.io/projected/0d7eb7b8-63ae-493a-850b-0b9f3b42e927-kube-api-access-dcbvc\") pod \"redhat-marketplace-g8bgt\" (UID: \"0d7eb7b8-63ae-493a-850b-0b9f3b42e927\") " pod="openshift-marketplace/redhat-marketplace-g8bgt" Jan 26 17:01:20 crc kubenswrapper[4856]: I0126 17:01:20.069059 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d7eb7b8-63ae-493a-850b-0b9f3b42e927-utilities\") pod \"redhat-marketplace-g8bgt\" (UID: \"0d7eb7b8-63ae-493a-850b-0b9f3b42e927\") " pod="openshift-marketplace/redhat-marketplace-g8bgt" Jan 26 17:01:20 crc kubenswrapper[4856]: I0126 17:01:20.069553 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d7eb7b8-63ae-493a-850b-0b9f3b42e927-utilities\") pod \"redhat-marketplace-g8bgt\" (UID: \"0d7eb7b8-63ae-493a-850b-0b9f3b42e927\") " pod="openshift-marketplace/redhat-marketplace-g8bgt" Jan 26 17:01:20 crc kubenswrapper[4856]: I0126 17:01:20.069786 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d7eb7b8-63ae-493a-850b-0b9f3b42e927-catalog-content\") pod \"redhat-marketplace-g8bgt\" (UID: \"0d7eb7b8-63ae-493a-850b-0b9f3b42e927\") " pod="openshift-marketplace/redhat-marketplace-g8bgt" Jan 26 17:01:20 crc kubenswrapper[4856]: I0126 17:01:20.117552 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcbvc\" (UniqueName: \"kubernetes.io/projected/0d7eb7b8-63ae-493a-850b-0b9f3b42e927-kube-api-access-dcbvc\") pod \"redhat-marketplace-g8bgt\" (UID: \"0d7eb7b8-63ae-493a-850b-0b9f3b42e927\") " pod="openshift-marketplace/redhat-marketplace-g8bgt" Jan 26 17:01:20 crc kubenswrapper[4856]: I0126 17:01:20.220289 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g8bgt" Jan 26 17:01:20 crc kubenswrapper[4856]: I0126 17:01:20.294087 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4kwt4"] Jan 26 17:01:20 crc kubenswrapper[4856]: W0126 17:01:20.369916 4856 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6944fc9_b8d7_4013_8702_b5765c410a0b.slice/crio-c6a85642ee783cdf59dd26ba744cc42773e760d42354900c16ebdd5e8e9ec111 WatchSource:0}: Error finding container c6a85642ee783cdf59dd26ba744cc42773e760d42354900c16ebdd5e8e9ec111: Status 404 returned error can't find the container with id c6a85642ee783cdf59dd26ba744cc42773e760d42354900c16ebdd5e8e9ec111 Jan 26 17:01:20 crc kubenswrapper[4856]: I0126 17:01:20.375993 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 26 17:01:20 crc kubenswrapper[4856]: I0126 17:01:20.474180 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d8c22047-144c-402a-80c5-c206539b6826-kube-api-access\") pod \"d8c22047-144c-402a-80c5-c206539b6826\" (UID: \"d8c22047-144c-402a-80c5-c206539b6826\") " Jan 26 17:01:20 crc kubenswrapper[4856]: I0126 17:01:20.474232 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d8c22047-144c-402a-80c5-c206539b6826-kubelet-dir\") pod \"d8c22047-144c-402a-80c5-c206539b6826\" (UID: \"d8c22047-144c-402a-80c5-c206539b6826\") " Jan 26 17:01:20 crc kubenswrapper[4856]: I0126 17:01:20.474734 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d8c22047-144c-402a-80c5-c206539b6826-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "d8c22047-144c-402a-80c5-c206539b6826" (UID: "d8c22047-144c-402a-80c5-c206539b6826"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:01:20 crc kubenswrapper[4856]: I0126 17:01:20.481782 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8c22047-144c-402a-80c5-c206539b6826-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "d8c22047-144c-402a-80c5-c206539b6826" (UID: "d8c22047-144c-402a-80c5-c206539b6826"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:01:20 crc kubenswrapper[4856]: I0126 17:01:20.576293 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d8c22047-144c-402a-80c5-c206539b6826-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 26 17:01:20 crc kubenswrapper[4856]: I0126 17:01:20.576329 4856 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d8c22047-144c-402a-80c5-c206539b6826-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 26 17:01:20 crc kubenswrapper[4856]: I0126 17:01:20.579981 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-g8bgt"] Jan 26 17:01:20 crc kubenswrapper[4856]: I0126 17:01:20.847003 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qknj9"] Jan 26 17:01:20 crc kubenswrapper[4856]: E0126 17:01:20.847607 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8c22047-144c-402a-80c5-c206539b6826" containerName="pruner" Jan 26 17:01:20 crc kubenswrapper[4856]: I0126 17:01:20.847625 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8c22047-144c-402a-80c5-c206539b6826" containerName="pruner" Jan 26 17:01:20 crc kubenswrapper[4856]: I0126 17:01:20.847797 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8c22047-144c-402a-80c5-c206539b6826" containerName="pruner" Jan 26 17:01:20 crc kubenswrapper[4856]: I0126 17:01:20.848515 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qknj9" Jan 26 17:01:20 crc kubenswrapper[4856]: I0126 17:01:20.853626 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 26 17:01:20 crc kubenswrapper[4856]: I0126 17:01:20.864201 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qknj9"] Jan 26 17:01:20 crc kubenswrapper[4856]: I0126 17:01:20.915420 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpqvc" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.006501 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3fa94fe-e4ad-4171-b853-89878dc61569-utilities\") pod \"redhat-operators-qknj9\" (UID: \"a3fa94fe-e4ad-4171-b853-89878dc61569\") " pod="openshift-marketplace/redhat-operators-qknj9" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.006691 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3fa94fe-e4ad-4171-b853-89878dc61569-catalog-content\") pod \"redhat-operators-qknj9\" (UID: \"a3fa94fe-e4ad-4171-b853-89878dc61569\") " pod="openshift-marketplace/redhat-operators-qknj9" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.006731 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvpzb\" (UniqueName: \"kubernetes.io/projected/a3fa94fe-e4ad-4171-b853-89878dc61569-kube-api-access-wvpzb\") pod \"redhat-operators-qknj9\" (UID: \"a3fa94fe-e4ad-4171-b853-89878dc61569\") " pod="openshift-marketplace/redhat-operators-qknj9" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.064789 4856 patch_prober.go:28] interesting pod/router-default-5444994796-h9b2g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 17:01:21 crc kubenswrapper[4856]: [-]has-synced failed: reason withheld Jan 26 17:01:21 crc kubenswrapper[4856]: [+]process-running ok Jan 26 17:01:21 crc kubenswrapper[4856]: healthz check failed Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.065199 4856 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9b2g" podUID="85f05bd5-ff83-4d29-9531-ab3499088095" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.080137 4856 generic.go:334] "Generic (PLEG): container finished" podID="d6944fc9-b8d7-4013-8702-b5765c410a0b" containerID="d4ffeb43e14865bfef28f884de6e5301087c2d9158d7a77b0c10a8dfec7c7ce2" exitCode=0 Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.080214 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4kwt4" event={"ID":"d6944fc9-b8d7-4013-8702-b5765c410a0b","Type":"ContainerDied","Data":"d4ffeb43e14865bfef28f884de6e5301087c2d9158d7a77b0c10a8dfec7c7ce2"} Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.080238 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4kwt4" event={"ID":"d6944fc9-b8d7-4013-8702-b5765c410a0b","Type":"ContainerStarted","Data":"c6a85642ee783cdf59dd26ba744cc42773e760d42354900c16ebdd5e8e9ec111"} Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.082000 4856 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.083151 4856 generic.go:334] "Generic (PLEG): container finished" podID="40a27476-22b1-4083-990e-66e70ccdaf4c" containerID="3ec09320bb48de5d8b6709469f0f84953408cf650f51d872373c21616d43f0de" exitCode=0 Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.083236 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-txmdl" event={"ID":"40a27476-22b1-4083-990e-66e70ccdaf4c","Type":"ContainerDied","Data":"3ec09320bb48de5d8b6709469f0f84953408cf650f51d872373c21616d43f0de"} Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.087194 4856 generic.go:334] "Generic (PLEG): container finished" podID="89cf05de-642b-4574-9f79-45e7a3d4afa3" containerID="de3e1fd7d5b6adab2150705e57df43577251e5278edb52956bb11f5539b1538a" exitCode=0 Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.087912 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qgjjd" event={"ID":"89cf05de-642b-4574-9f79-45e7a3d4afa3","Type":"ContainerDied","Data":"de3e1fd7d5b6adab2150705e57df43577251e5278edb52956bb11f5539b1538a"} Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.095240 4856 generic.go:334] "Generic (PLEG): container finished" podID="0d7eb7b8-63ae-493a-850b-0b9f3b42e927" containerID="a9fe692a78995f7dad7ea556edacc772eb429ab92938195725add9a17bbe9e7c" exitCode=0 Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.095422 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g8bgt" event={"ID":"0d7eb7b8-63ae-493a-850b-0b9f3b42e927","Type":"ContainerDied","Data":"a9fe692a78995f7dad7ea556edacc772eb429ab92938195725add9a17bbe9e7c"} Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.096558 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g8bgt" event={"ID":"0d7eb7b8-63ae-493a-850b-0b9f3b42e927","Type":"ContainerStarted","Data":"eca9c93c5c35ce3c6c300c833124d2e0c4c40f4feaf4a45bd12b4eecdb2f116c"} Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.098845 4856 generic.go:334] "Generic (PLEG): container finished" podID="5fe6baed-ab97-4d8a-8be2-6f00f9698136" containerID="56133c3e036efeb9590dc043f9b9af766fce603e2c50cfdca46be37466b88f62" exitCode=0 Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.098918 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpqvc" event={"ID":"5fe6baed-ab97-4d8a-8be2-6f00f9698136","Type":"ContainerDied","Data":"56133c3e036efeb9590dc043f9b9af766fce603e2c50cfdca46be37466b88f62"} Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.098960 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpqvc" event={"ID":"5fe6baed-ab97-4d8a-8be2-6f00f9698136","Type":"ContainerDied","Data":"e3a4f0c156036789efac8b4cdbd3ace5dcdaf8c187d261687c3b9c87a15d74df"} Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.099000 4856 scope.go:117] "RemoveContainer" containerID="56133c3e036efeb9590dc043f9b9af766fce603e2c50cfdca46be37466b88f62" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.099121 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpqvc" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.107173 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fe6baed-ab97-4d8a-8be2-6f00f9698136-config\") pod \"5fe6baed-ab97-4d8a-8be2-6f00f9698136\" (UID: \"5fe6baed-ab97-4d8a-8be2-6f00f9698136\") " Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.107214 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5fe6baed-ab97-4d8a-8be2-6f00f9698136-client-ca\") pod \"5fe6baed-ab97-4d8a-8be2-6f00f9698136\" (UID: \"5fe6baed-ab97-4d8a-8be2-6f00f9698136\") " Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.107240 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hpfwk\" (UniqueName: \"kubernetes.io/projected/5fe6baed-ab97-4d8a-8be2-6f00f9698136-kube-api-access-hpfwk\") pod \"5fe6baed-ab97-4d8a-8be2-6f00f9698136\" (UID: \"5fe6baed-ab97-4d8a-8be2-6f00f9698136\") " Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.107308 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5fe6baed-ab97-4d8a-8be2-6f00f9698136-serving-cert\") pod \"5fe6baed-ab97-4d8a-8be2-6f00f9698136\" (UID: \"5fe6baed-ab97-4d8a-8be2-6f00f9698136\") " Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.107541 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3fa94fe-e4ad-4171-b853-89878dc61569-catalog-content\") pod \"redhat-operators-qknj9\" (UID: \"a3fa94fe-e4ad-4171-b853-89878dc61569\") " pod="openshift-marketplace/redhat-operators-qknj9" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.107568 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvpzb\" (UniqueName: \"kubernetes.io/projected/a3fa94fe-e4ad-4171-b853-89878dc61569-kube-api-access-wvpzb\") pod \"redhat-operators-qknj9\" (UID: \"a3fa94fe-e4ad-4171-b853-89878dc61569\") " pod="openshift-marketplace/redhat-operators-qknj9" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.107587 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3fa94fe-e4ad-4171-b853-89878dc61569-utilities\") pod \"redhat-operators-qknj9\" (UID: \"a3fa94fe-e4ad-4171-b853-89878dc61569\") " pod="openshift-marketplace/redhat-operators-qknj9" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.108038 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3fa94fe-e4ad-4171-b853-89878dc61569-utilities\") pod \"redhat-operators-qknj9\" (UID: \"a3fa94fe-e4ad-4171-b853-89878dc61569\") " pod="openshift-marketplace/redhat-operators-qknj9" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.110265 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5fe6baed-ab97-4d8a-8be2-6f00f9698136-config" (OuterVolumeSpecName: "config") pod "5fe6baed-ab97-4d8a-8be2-6f00f9698136" (UID: "5fe6baed-ab97-4d8a-8be2-6f00f9698136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.110711 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5fe6baed-ab97-4d8a-8be2-6f00f9698136-client-ca" (OuterVolumeSpecName: "client-ca") pod "5fe6baed-ab97-4d8a-8be2-6f00f9698136" (UID: "5fe6baed-ab97-4d8a-8be2-6f00f9698136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.111380 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3fa94fe-e4ad-4171-b853-89878dc61569-catalog-content\") pod \"redhat-operators-qknj9\" (UID: \"a3fa94fe-e4ad-4171-b853-89878dc61569\") " pod="openshift-marketplace/redhat-operators-qknj9" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.113515 4856 generic.go:334] "Generic (PLEG): container finished" podID="a6086d4b-faeb-4a12-8e6a-2a178dfe374c" containerID="5638f22e046bc8f28ee2834fa7820e942af58e17d4efe952168ca98e63b3fa12" exitCode=0 Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.113668 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n8hp2" event={"ID":"a6086d4b-faeb-4a12-8e6a-2a178dfe374c","Type":"ContainerDied","Data":"5638f22e046bc8f28ee2834fa7820e942af58e17d4efe952168ca98e63b3fa12"} Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.117039 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-295wr" event={"ID":"12e50462-28e6-4531-ada4-e652310e6cce","Type":"ContainerStarted","Data":"5868fcdab59505042d1235014aa1685fbf9cc65d23b45066a7b3729afc514f50"} Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.117088 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-295wr" event={"ID":"12e50462-28e6-4531-ada4-e652310e6cce","Type":"ContainerStarted","Data":"7de6d5aef139d7d3e970838bf0c7a91c17246160a4fbf31d318d10f88ebf2901"} Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.118690 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe6baed-ab97-4d8a-8be2-6f00f9698136-kube-api-access-hpfwk" (OuterVolumeSpecName: "kube-api-access-hpfwk") pod "5fe6baed-ab97-4d8a-8be2-6f00f9698136" (UID: "5fe6baed-ab97-4d8a-8be2-6f00f9698136"). InnerVolumeSpecName "kube-api-access-hpfwk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.119569 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe6baed-ab97-4d8a-8be2-6f00f9698136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5fe6baed-ab97-4d8a-8be2-6f00f9698136" (UID: "5fe6baed-ab97-4d8a-8be2-6f00f9698136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.122135 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" event={"ID":"cfa40861-cc08-4145-a185-6a3fb07eaabe","Type":"ContainerStarted","Data":"fc8e05e1e87fe66232302aff71c23d6b6c36b366751f113f41815a46bc948eb9"} Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.122171 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" event={"ID":"cfa40861-cc08-4145-a185-6a3fb07eaabe","Type":"ContainerStarted","Data":"ae7df2de181ac684cadd8c52c3b8878c72703f16549d24e92a2fc45b186ce717"} Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.122267 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.129994 4856 generic.go:334] "Generic (PLEG): container finished" podID="1afc0f4c-e02d-4a70-aaba-e761e8c04eee" containerID="e9e54e2a4a2266ca4148b11cb38df08f87c2f2ccd87dc3343d147862786c16e2" exitCode=0 Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.130121 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-lndnt" event={"ID":"1afc0f4c-e02d-4a70-aaba-e761e8c04eee","Type":"ContainerDied","Data":"e9e54e2a4a2266ca4148b11cb38df08f87c2f2ccd87dc3343d147862786c16e2"} Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.133216 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"d8c22047-144c-402a-80c5-c206539b6826","Type":"ContainerDied","Data":"eb80e342e945ab720a056481e6d379636786d87ebb81ecfa7bcd84ffb36388ff"} Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.133245 4856 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb80e342e945ab720a056481e6d379636786d87ebb81ecfa7bcd84ffb36388ff" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.133308 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.143611 4856 generic.go:334] "Generic (PLEG): container finished" podID="7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766" containerID="f96de8f882682ea8e5a30970c1ce8d34c4b60cb434e13968e3bd6879b62b071b" exitCode=0 Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.143669 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-62nhd" event={"ID":"7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766","Type":"ContainerDied","Data":"f96de8f882682ea8e5a30970c1ce8d34c4b60cb434e13968e3bd6879b62b071b"} Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.147461 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvpzb\" (UniqueName: \"kubernetes.io/projected/a3fa94fe-e4ad-4171-b853-89878dc61569-kube-api-access-wvpzb\") pod \"redhat-operators-qknj9\" (UID: \"a3fa94fe-e4ad-4171-b853-89878dc61569\") " pod="openshift-marketplace/redhat-operators-qknj9" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.158962 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-lndnt" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.160304 4856 scope.go:117] "RemoveContainer" containerID="56133c3e036efeb9590dc043f9b9af766fce603e2c50cfdca46be37466b88f62" Jan 26 17:01:21 crc kubenswrapper[4856]: E0126 17:01:21.160654 4856 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56133c3e036efeb9590dc043f9b9af766fce603e2c50cfdca46be37466b88f62\": container with ID starting with 56133c3e036efeb9590dc043f9b9af766fce603e2c50cfdca46be37466b88f62 not found: ID does not exist" containerID="56133c3e036efeb9590dc043f9b9af766fce603e2c50cfdca46be37466b88f62" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.160694 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56133c3e036efeb9590dc043f9b9af766fce603e2c50cfdca46be37466b88f62"} err="failed to get container status \"56133c3e036efeb9590dc043f9b9af766fce603e2c50cfdca46be37466b88f62\": rpc error: code = NotFound desc = could not find container \"56133c3e036efeb9590dc043f9b9af766fce603e2c50cfdca46be37466b88f62\": container with ID starting with 56133c3e036efeb9590dc043f9b9af766fce603e2c50cfdca46be37466b88f62 not found: ID does not exist" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.209038 4856 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fe6baed-ab97-4d8a-8be2-6f00f9698136-config\") on node \"crc\" DevicePath \"\"" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.209076 4856 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5fe6baed-ab97-4d8a-8be2-6f00f9698136-client-ca\") on node \"crc\" DevicePath \"\"" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.209088 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hpfwk\" (UniqueName: \"kubernetes.io/projected/5fe6baed-ab97-4d8a-8be2-6f00f9698136-kube-api-access-hpfwk\") on node \"crc\" DevicePath \"\"" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.209100 4856 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5fe6baed-ab97-4d8a-8be2-6f00f9698136-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.228805 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qknj9" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.261128 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5457c9f8cf-lcqcl"] Jan 26 17:01:21 crc kubenswrapper[4856]: E0126 17:01:21.261364 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fe6baed-ab97-4d8a-8be2-6f00f9698136" containerName="route-controller-manager" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.261375 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fe6baed-ab97-4d8a-8be2-6f00f9698136" containerName="route-controller-manager" Jan 26 17:01:21 crc kubenswrapper[4856]: E0126 17:01:21.261391 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1afc0f4c-e02d-4a70-aaba-e761e8c04eee" containerName="controller-manager" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.261397 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="1afc0f4c-e02d-4a70-aaba-e761e8c04eee" containerName="controller-manager" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.261501 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fe6baed-ab97-4d8a-8be2-6f00f9698136" containerName="route-controller-manager" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.261514 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="1afc0f4c-e02d-4a70-aaba-e761e8c04eee" containerName="controller-manager" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.261854 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cb449784d-bqprm"] Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.271877 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-cb449784d-bqprm" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.272329 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5457c9f8cf-lcqcl" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.304377 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mqxwf"] Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.310992 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mqxwf" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.313626 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1afc0f4c-e02d-4a70-aaba-e761e8c04eee-client-ca\") pod \"1afc0f4c-e02d-4a70-aaba-e761e8c04eee\" (UID: \"1afc0f4c-e02d-4a70-aaba-e761e8c04eee\") " Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.314206 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mr8gn\" (UniqueName: \"kubernetes.io/projected/1afc0f4c-e02d-4a70-aaba-e761e8c04eee-kube-api-access-mr8gn\") pod \"1afc0f4c-e02d-4a70-aaba-e761e8c04eee\" (UID: \"1afc0f4c-e02d-4a70-aaba-e761e8c04eee\") " Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.314250 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1afc0f4c-e02d-4a70-aaba-e761e8c04eee-proxy-ca-bundles\") pod \"1afc0f4c-e02d-4a70-aaba-e761e8c04eee\" (UID: \"1afc0f4c-e02d-4a70-aaba-e761e8c04eee\") " Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.315155 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1afc0f4c-e02d-4a70-aaba-e761e8c04eee-serving-cert\") pod \"1afc0f4c-e02d-4a70-aaba-e761e8c04eee\" (UID: \"1afc0f4c-e02d-4a70-aaba-e761e8c04eee\") " Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.315654 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1afc0f4c-e02d-4a70-aaba-e761e8c04eee-config\") pod \"1afc0f4c-e02d-4a70-aaba-e761e8c04eee\" (UID: \"1afc0f4c-e02d-4a70-aaba-e761e8c04eee\") " Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.316270 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-295wr" podStartSLOduration=147.316251539 podStartE2EDuration="2m27.316251539s" podCreationTimestamp="2026-01-26 16:58:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:01:21.273159958 +0000 UTC m=+177.226413939" watchObservedRunningTime="2026-01-26 17:01:21.316251539 +0000 UTC m=+177.269505520" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.319585 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1afc0f4c-e02d-4a70-aaba-e761e8c04eee-client-ca" (OuterVolumeSpecName: "client-ca") pod "1afc0f4c-e02d-4a70-aaba-e761e8c04eee" (UID: "1afc0f4c-e02d-4a70-aaba-e761e8c04eee"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.321073 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1afc0f4c-e02d-4a70-aaba-e761e8c04eee-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "1afc0f4c-e02d-4a70-aaba-e761e8c04eee" (UID: "1afc0f4c-e02d-4a70-aaba-e761e8c04eee"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.322181 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1afc0f4c-e02d-4a70-aaba-e761e8c04eee-config" (OuterVolumeSpecName: "config") pod "1afc0f4c-e02d-4a70-aaba-e761e8c04eee" (UID: "1afc0f4c-e02d-4a70-aaba-e761e8c04eee"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.334057 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1afc0f4c-e02d-4a70-aaba-e761e8c04eee-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1afc0f4c-e02d-4a70-aaba-e761e8c04eee" (UID: "1afc0f4c-e02d-4a70-aaba-e761e8c04eee"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.339691 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5457c9f8cf-lcqcl"] Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.351919 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mqxwf"] Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.357645 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1afc0f4c-e02d-4a70-aaba-e761e8c04eee-kube-api-access-mr8gn" (OuterVolumeSpecName: "kube-api-access-mr8gn") pod "1afc0f4c-e02d-4a70-aaba-e761e8c04eee" (UID: "1afc0f4c-e02d-4a70-aaba-e761e8c04eee"). InnerVolumeSpecName "kube-api-access-mr8gn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.362133 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" podStartSLOduration=147.362102271 podStartE2EDuration="2m27.362102271s" podCreationTimestamp="2026-01-26 16:58:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:01:21.333028304 +0000 UTC m=+177.286282285" watchObservedRunningTime="2026-01-26 17:01:21.362102271 +0000 UTC m=+177.315356262" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.396599 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cb449784d-bqprm"] Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.420604 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vhl4\" (UniqueName: \"kubernetes.io/projected/54dee8cd-259a-4c9f-9e56-fbd0ea167f46-kube-api-access-9vhl4\") pod \"route-controller-manager-cb449784d-bqprm\" (UID: \"54dee8cd-259a-4c9f-9e56-fbd0ea167f46\") " pod="openshift-route-controller-manager/route-controller-manager-cb449784d-bqprm" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.420649 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c71e219-35d7-4e1e-a371-3456dfd29e83-utilities\") pod \"redhat-operators-mqxwf\" (UID: \"9c71e219-35d7-4e1e-a371-3456dfd29e83\") " pod="openshift-marketplace/redhat-operators-mqxwf" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.420671 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4sjtb\" (UniqueName: \"kubernetes.io/projected/50b63435-18c4-4fcb-821c-3d88abc7b728-kube-api-access-4sjtb\") pod \"controller-manager-5457c9f8cf-lcqcl\" (UID: \"50b63435-18c4-4fcb-821c-3d88abc7b728\") " pod="openshift-controller-manager/controller-manager-5457c9f8cf-lcqcl" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.420694 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c71e219-35d7-4e1e-a371-3456dfd29e83-catalog-content\") pod \"redhat-operators-mqxwf\" (UID: \"9c71e219-35d7-4e1e-a371-3456dfd29e83\") " pod="openshift-marketplace/redhat-operators-mqxwf" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.420713 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6swzw\" (UniqueName: \"kubernetes.io/projected/9c71e219-35d7-4e1e-a371-3456dfd29e83-kube-api-access-6swzw\") pod \"redhat-operators-mqxwf\" (UID: \"9c71e219-35d7-4e1e-a371-3456dfd29e83\") " pod="openshift-marketplace/redhat-operators-mqxwf" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.420742 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/50b63435-18c4-4fcb-821c-3d88abc7b728-proxy-ca-bundles\") pod \"controller-manager-5457c9f8cf-lcqcl\" (UID: \"50b63435-18c4-4fcb-821c-3d88abc7b728\") " pod="openshift-controller-manager/controller-manager-5457c9f8cf-lcqcl" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.420769 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/54dee8cd-259a-4c9f-9e56-fbd0ea167f46-client-ca\") pod \"route-controller-manager-cb449784d-bqprm\" (UID: \"54dee8cd-259a-4c9f-9e56-fbd0ea167f46\") " pod="openshift-route-controller-manager/route-controller-manager-cb449784d-bqprm" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.420791 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50b63435-18c4-4fcb-821c-3d88abc7b728-serving-cert\") pod \"controller-manager-5457c9f8cf-lcqcl\" (UID: \"50b63435-18c4-4fcb-821c-3d88abc7b728\") " pod="openshift-controller-manager/controller-manager-5457c9f8cf-lcqcl" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.420810 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54dee8cd-259a-4c9f-9e56-fbd0ea167f46-serving-cert\") pod \"route-controller-manager-cb449784d-bqprm\" (UID: \"54dee8cd-259a-4c9f-9e56-fbd0ea167f46\") " pod="openshift-route-controller-manager/route-controller-manager-cb449784d-bqprm" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.420835 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/50b63435-18c4-4fcb-821c-3d88abc7b728-client-ca\") pod \"controller-manager-5457c9f8cf-lcqcl\" (UID: \"50b63435-18c4-4fcb-821c-3d88abc7b728\") " pod="openshift-controller-manager/controller-manager-5457c9f8cf-lcqcl" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.420867 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54dee8cd-259a-4c9f-9e56-fbd0ea167f46-config\") pod \"route-controller-manager-cb449784d-bqprm\" (UID: \"54dee8cd-259a-4c9f-9e56-fbd0ea167f46\") " pod="openshift-route-controller-manager/route-controller-manager-cb449784d-bqprm" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.420883 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/50b63435-18c4-4fcb-821c-3d88abc7b728-config\") pod \"controller-manager-5457c9f8cf-lcqcl\" (UID: \"50b63435-18c4-4fcb-821c-3d88abc7b728\") " pod="openshift-controller-manager/controller-manager-5457c9f8cf-lcqcl" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.420916 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mr8gn\" (UniqueName: \"kubernetes.io/projected/1afc0f4c-e02d-4a70-aaba-e761e8c04eee-kube-api-access-mr8gn\") on node \"crc\" DevicePath \"\"" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.420927 4856 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1afc0f4c-e02d-4a70-aaba-e761e8c04eee-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.420938 4856 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1afc0f4c-e02d-4a70-aaba-e761e8c04eee-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.420947 4856 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1afc0f4c-e02d-4a70-aaba-e761e8c04eee-config\") on node \"crc\" DevicePath \"\"" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.420955 4856 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1afc0f4c-e02d-4a70-aaba-e761e8c04eee-client-ca\") on node \"crc\" DevicePath \"\"" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.428998 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.429842 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5457c9f8cf-lcqcl"] Jan 26 17:01:21 crc kubenswrapper[4856]: E0126 17:01:21.430194 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[client-ca config kube-api-access-4sjtb proxy-ca-bundles serving-cert], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-controller-manager/controller-manager-5457c9f8cf-lcqcl" podUID="50b63435-18c4-4fcb-821c-3d88abc7b728" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.488176 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpqvc"] Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.492891 4856 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-fpqvc"] Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.535395 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50b63435-18c4-4fcb-821c-3d88abc7b728-serving-cert\") pod \"controller-manager-5457c9f8cf-lcqcl\" (UID: \"50b63435-18c4-4fcb-821c-3d88abc7b728\") " pod="openshift-controller-manager/controller-manager-5457c9f8cf-lcqcl" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.535448 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54dee8cd-259a-4c9f-9e56-fbd0ea167f46-serving-cert\") pod \"route-controller-manager-cb449784d-bqprm\" (UID: \"54dee8cd-259a-4c9f-9e56-fbd0ea167f46\") " pod="openshift-route-controller-manager/route-controller-manager-cb449784d-bqprm" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.535482 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/50b63435-18c4-4fcb-821c-3d88abc7b728-client-ca\") pod \"controller-manager-5457c9f8cf-lcqcl\" (UID: \"50b63435-18c4-4fcb-821c-3d88abc7b728\") " pod="openshift-controller-manager/controller-manager-5457c9f8cf-lcqcl" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.535547 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54dee8cd-259a-4c9f-9e56-fbd0ea167f46-config\") pod \"route-controller-manager-cb449784d-bqprm\" (UID: \"54dee8cd-259a-4c9f-9e56-fbd0ea167f46\") " pod="openshift-route-controller-manager/route-controller-manager-cb449784d-bqprm" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.535586 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/50b63435-18c4-4fcb-821c-3d88abc7b728-config\") pod \"controller-manager-5457c9f8cf-lcqcl\" (UID: \"50b63435-18c4-4fcb-821c-3d88abc7b728\") " pod="openshift-controller-manager/controller-manager-5457c9f8cf-lcqcl" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.535630 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vhl4\" (UniqueName: \"kubernetes.io/projected/54dee8cd-259a-4c9f-9e56-fbd0ea167f46-kube-api-access-9vhl4\") pod \"route-controller-manager-cb449784d-bqprm\" (UID: \"54dee8cd-259a-4c9f-9e56-fbd0ea167f46\") " pod="openshift-route-controller-manager/route-controller-manager-cb449784d-bqprm" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.535664 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c71e219-35d7-4e1e-a371-3456dfd29e83-utilities\") pod \"redhat-operators-mqxwf\" (UID: \"9c71e219-35d7-4e1e-a371-3456dfd29e83\") " pod="openshift-marketplace/redhat-operators-mqxwf" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.535686 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4sjtb\" (UniqueName: \"kubernetes.io/projected/50b63435-18c4-4fcb-821c-3d88abc7b728-kube-api-access-4sjtb\") pod \"controller-manager-5457c9f8cf-lcqcl\" (UID: \"50b63435-18c4-4fcb-821c-3d88abc7b728\") " pod="openshift-controller-manager/controller-manager-5457c9f8cf-lcqcl" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.535709 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c71e219-35d7-4e1e-a371-3456dfd29e83-catalog-content\") pod \"redhat-operators-mqxwf\" (UID: \"9c71e219-35d7-4e1e-a371-3456dfd29e83\") " pod="openshift-marketplace/redhat-operators-mqxwf" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.535727 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6swzw\" (UniqueName: \"kubernetes.io/projected/9c71e219-35d7-4e1e-a371-3456dfd29e83-kube-api-access-6swzw\") pod \"redhat-operators-mqxwf\" (UID: \"9c71e219-35d7-4e1e-a371-3456dfd29e83\") " pod="openshift-marketplace/redhat-operators-mqxwf" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.535754 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/50b63435-18c4-4fcb-821c-3d88abc7b728-proxy-ca-bundles\") pod \"controller-manager-5457c9f8cf-lcqcl\" (UID: \"50b63435-18c4-4fcb-821c-3d88abc7b728\") " pod="openshift-controller-manager/controller-manager-5457c9f8cf-lcqcl" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.535781 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/54dee8cd-259a-4c9f-9e56-fbd0ea167f46-client-ca\") pod \"route-controller-manager-cb449784d-bqprm\" (UID: \"54dee8cd-259a-4c9f-9e56-fbd0ea167f46\") " pod="openshift-route-controller-manager/route-controller-manager-cb449784d-bqprm" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.536690 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c71e219-35d7-4e1e-a371-3456dfd29e83-utilities\") pod \"redhat-operators-mqxwf\" (UID: \"9c71e219-35d7-4e1e-a371-3456dfd29e83\") " pod="openshift-marketplace/redhat-operators-mqxwf" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.536960 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c71e219-35d7-4e1e-a371-3456dfd29e83-catalog-content\") pod \"redhat-operators-mqxwf\" (UID: \"9c71e219-35d7-4e1e-a371-3456dfd29e83\") " pod="openshift-marketplace/redhat-operators-mqxwf" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.537564 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/50b63435-18c4-4fcb-821c-3d88abc7b728-client-ca\") pod \"controller-manager-5457c9f8cf-lcqcl\" (UID: \"50b63435-18c4-4fcb-821c-3d88abc7b728\") " pod="openshift-controller-manager/controller-manager-5457c9f8cf-lcqcl" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.538491 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/50b63435-18c4-4fcb-821c-3d88abc7b728-config\") pod \"controller-manager-5457c9f8cf-lcqcl\" (UID: \"50b63435-18c4-4fcb-821c-3d88abc7b728\") " pod="openshift-controller-manager/controller-manager-5457c9f8cf-lcqcl" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.540582 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54dee8cd-259a-4c9f-9e56-fbd0ea167f46-config\") pod \"route-controller-manager-cb449784d-bqprm\" (UID: \"54dee8cd-259a-4c9f-9e56-fbd0ea167f46\") " pod="openshift-route-controller-manager/route-controller-manager-cb449784d-bqprm" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.540939 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54dee8cd-259a-4c9f-9e56-fbd0ea167f46-serving-cert\") pod \"route-controller-manager-cb449784d-bqprm\" (UID: \"54dee8cd-259a-4c9f-9e56-fbd0ea167f46\") " pod="openshift-route-controller-manager/route-controller-manager-cb449784d-bqprm" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.543100 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50b63435-18c4-4fcb-821c-3d88abc7b728-serving-cert\") pod \"controller-manager-5457c9f8cf-lcqcl\" (UID: \"50b63435-18c4-4fcb-821c-3d88abc7b728\") " pod="openshift-controller-manager/controller-manager-5457c9f8cf-lcqcl" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.543730 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/50b63435-18c4-4fcb-821c-3d88abc7b728-proxy-ca-bundles\") pod \"controller-manager-5457c9f8cf-lcqcl\" (UID: \"50b63435-18c4-4fcb-821c-3d88abc7b728\") " pod="openshift-controller-manager/controller-manager-5457c9f8cf-lcqcl" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.554024 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6swzw\" (UniqueName: \"kubernetes.io/projected/9c71e219-35d7-4e1e-a371-3456dfd29e83-kube-api-access-6swzw\") pod \"redhat-operators-mqxwf\" (UID: \"9c71e219-35d7-4e1e-a371-3456dfd29e83\") " pod="openshift-marketplace/redhat-operators-mqxwf" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.556991 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/54dee8cd-259a-4c9f-9e56-fbd0ea167f46-client-ca\") pod \"route-controller-manager-cb449784d-bqprm\" (UID: \"54dee8cd-259a-4c9f-9e56-fbd0ea167f46\") " pod="openshift-route-controller-manager/route-controller-manager-cb449784d-bqprm" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.557843 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vhl4\" (UniqueName: \"kubernetes.io/projected/54dee8cd-259a-4c9f-9e56-fbd0ea167f46-kube-api-access-9vhl4\") pod \"route-controller-manager-cb449784d-bqprm\" (UID: \"54dee8cd-259a-4c9f-9e56-fbd0ea167f46\") " pod="openshift-route-controller-manager/route-controller-manager-cb449784d-bqprm" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.559242 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4sjtb\" (UniqueName: \"kubernetes.io/projected/50b63435-18c4-4fcb-821c-3d88abc7b728-kube-api-access-4sjtb\") pod \"controller-manager-5457c9f8cf-lcqcl\" (UID: \"50b63435-18c4-4fcb-821c-3d88abc7b728\") " pod="openshift-controller-manager/controller-manager-5457c9f8cf-lcqcl" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.599149 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qknj9"] Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.616106 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-cb449784d-bqprm" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.647485 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mqxwf" Jan 26 17:01:21 crc kubenswrapper[4856]: I0126 17:01:21.906266 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mqxwf"] Jan 26 17:01:21 crc kubenswrapper[4856]: W0126 17:01:21.971695 4856 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9c71e219_35d7_4e1e_a371_3456dfd29e83.slice/crio-d09c4604a24ed1fd63afc114569ecaa6c0c08542e351c04817bb0f8a62c19b49 WatchSource:0}: Error finding container d09c4604a24ed1fd63afc114569ecaa6c0c08542e351c04817bb0f8a62c19b49: Status 404 returned error can't find the container with id d09c4604a24ed1fd63afc114569ecaa6c0c08542e351c04817bb0f8a62c19b49 Jan 26 17:01:22 crc kubenswrapper[4856]: I0126 17:01:22.059682 4856 patch_prober.go:28] interesting pod/router-default-5444994796-h9b2g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 17:01:22 crc kubenswrapper[4856]: [-]has-synced failed: reason withheld Jan 26 17:01:22 crc kubenswrapper[4856]: [+]process-running ok Jan 26 17:01:22 crc kubenswrapper[4856]: healthz check failed Jan 26 17:01:22 crc kubenswrapper[4856]: I0126 17:01:22.059738 4856 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9b2g" podUID="85f05bd5-ff83-4d29-9531-ab3499088095" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 17:01:22 crc kubenswrapper[4856]: I0126 17:01:22.155413 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cb449784d-bqprm"] Jan 26 17:01:22 crc kubenswrapper[4856]: I0126 17:01:22.157612 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mqxwf" event={"ID":"9c71e219-35d7-4e1e-a371-3456dfd29e83","Type":"ContainerStarted","Data":"e6e9fc1c7474ee1cf14a50a96e79036f97d946d338d8a18c3434197cbd0438a8"} Jan 26 17:01:22 crc kubenswrapper[4856]: I0126 17:01:22.157653 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mqxwf" event={"ID":"9c71e219-35d7-4e1e-a371-3456dfd29e83","Type":"ContainerStarted","Data":"d09c4604a24ed1fd63afc114569ecaa6c0c08542e351c04817bb0f8a62c19b49"} Jan 26 17:01:22 crc kubenswrapper[4856]: I0126 17:01:22.160312 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-lndnt" event={"ID":"1afc0f4c-e02d-4a70-aaba-e761e8c04eee","Type":"ContainerDied","Data":"96859d6a59b58c9df792a590deef50eb0ee923d03cb16fdc72abe3d18e466eaa"} Jan 26 17:01:22 crc kubenswrapper[4856]: I0126 17:01:22.160907 4856 scope.go:117] "RemoveContainer" containerID="e9e54e2a4a2266ca4148b11cb38df08f87c2f2ccd87dc3343d147862786c16e2" Jan 26 17:01:22 crc kubenswrapper[4856]: I0126 17:01:22.160324 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-lndnt" Jan 26 17:01:22 crc kubenswrapper[4856]: I0126 17:01:22.164206 4856 generic.go:334] "Generic (PLEG): container finished" podID="a3fa94fe-e4ad-4171-b853-89878dc61569" containerID="6c718aeedef34f07c2686370f8f78fe4060881e116396cc02bb806370cffdb47" exitCode=0 Jan 26 17:01:22 crc kubenswrapper[4856]: I0126 17:01:22.164308 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qknj9" event={"ID":"a3fa94fe-e4ad-4171-b853-89878dc61569","Type":"ContainerDied","Data":"6c718aeedef34f07c2686370f8f78fe4060881e116396cc02bb806370cffdb47"} Jan 26 17:01:22 crc kubenswrapper[4856]: I0126 17:01:22.164342 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qknj9" event={"ID":"a3fa94fe-e4ad-4171-b853-89878dc61569","Type":"ContainerStarted","Data":"083b0b52d78f857657f62965a6b3636eba0ff933ac74b23de919043206cf9046"} Jan 26 17:01:22 crc kubenswrapper[4856]: I0126 17:01:22.171479 4856 generic.go:334] "Generic (PLEG): container finished" podID="7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7" containerID="655c350d2621ac99cae47d6117abe996be96564e1734dccd0a74e6f8446d8e6d" exitCode=0 Jan 26 17:01:22 crc kubenswrapper[4856]: I0126 17:01:22.172354 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-8q6q4" event={"ID":"7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7","Type":"ContainerDied","Data":"655c350d2621ac99cae47d6117abe996be96564e1734dccd0a74e6f8446d8e6d"} Jan 26 17:01:22 crc kubenswrapper[4856]: I0126 17:01:22.172493 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5457c9f8cf-lcqcl" Jan 26 17:01:22 crc kubenswrapper[4856]: W0126 17:01:22.201807 4856 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod54dee8cd_259a_4c9f_9e56_fbd0ea167f46.slice/crio-960ff00bedf28636eb04c4f352e2d6d2e33a5ceb9800e901e018103cd5ac5859 WatchSource:0}: Error finding container 960ff00bedf28636eb04c4f352e2d6d2e33a5ceb9800e901e018103cd5ac5859: Status 404 returned error can't find the container with id 960ff00bedf28636eb04c4f352e2d6d2e33a5ceb9800e901e018103cd5ac5859 Jan 26 17:01:22 crc kubenswrapper[4856]: I0126 17:01:22.219045 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5457c9f8cf-lcqcl" Jan 26 17:01:22 crc kubenswrapper[4856]: I0126 17:01:22.236418 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lndnt"] Jan 26 17:01:22 crc kubenswrapper[4856]: I0126 17:01:22.240339 4856 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-lndnt"] Jan 26 17:01:22 crc kubenswrapper[4856]: I0126 17:01:22.348003 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/50b63435-18c4-4fcb-821c-3d88abc7b728-client-ca\") pod \"50b63435-18c4-4fcb-821c-3d88abc7b728\" (UID: \"50b63435-18c4-4fcb-821c-3d88abc7b728\") " Jan 26 17:01:22 crc kubenswrapper[4856]: I0126 17:01:22.348636 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4sjtb\" (UniqueName: \"kubernetes.io/projected/50b63435-18c4-4fcb-821c-3d88abc7b728-kube-api-access-4sjtb\") pod \"50b63435-18c4-4fcb-821c-3d88abc7b728\" (UID: \"50b63435-18c4-4fcb-821c-3d88abc7b728\") " Jan 26 17:01:22 crc kubenswrapper[4856]: I0126 17:01:22.348731 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/50b63435-18c4-4fcb-821c-3d88abc7b728-proxy-ca-bundles\") pod \"50b63435-18c4-4fcb-821c-3d88abc7b728\" (UID: \"50b63435-18c4-4fcb-821c-3d88abc7b728\") " Jan 26 17:01:22 crc kubenswrapper[4856]: I0126 17:01:22.348750 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/50b63435-18c4-4fcb-821c-3d88abc7b728-config\") pod \"50b63435-18c4-4fcb-821c-3d88abc7b728\" (UID: \"50b63435-18c4-4fcb-821c-3d88abc7b728\") " Jan 26 17:01:22 crc kubenswrapper[4856]: I0126 17:01:22.348730 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/50b63435-18c4-4fcb-821c-3d88abc7b728-client-ca" (OuterVolumeSpecName: "client-ca") pod "50b63435-18c4-4fcb-821c-3d88abc7b728" (UID: "50b63435-18c4-4fcb-821c-3d88abc7b728"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:01:22 crc kubenswrapper[4856]: I0126 17:01:22.348771 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50b63435-18c4-4fcb-821c-3d88abc7b728-serving-cert\") pod \"50b63435-18c4-4fcb-821c-3d88abc7b728\" (UID: \"50b63435-18c4-4fcb-821c-3d88abc7b728\") " Jan 26 17:01:22 crc kubenswrapper[4856]: I0126 17:01:22.349317 4856 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/50b63435-18c4-4fcb-821c-3d88abc7b728-client-ca\") on node \"crc\" DevicePath \"\"" Jan 26 17:01:22 crc kubenswrapper[4856]: I0126 17:01:22.349333 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/50b63435-18c4-4fcb-821c-3d88abc7b728-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "50b63435-18c4-4fcb-821c-3d88abc7b728" (UID: "50b63435-18c4-4fcb-821c-3d88abc7b728"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:01:22 crc kubenswrapper[4856]: I0126 17:01:22.352259 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/50b63435-18c4-4fcb-821c-3d88abc7b728-config" (OuterVolumeSpecName: "config") pod "50b63435-18c4-4fcb-821c-3d88abc7b728" (UID: "50b63435-18c4-4fcb-821c-3d88abc7b728"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:01:22 crc kubenswrapper[4856]: I0126 17:01:22.364718 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50b63435-18c4-4fcb-821c-3d88abc7b728-kube-api-access-4sjtb" (OuterVolumeSpecName: "kube-api-access-4sjtb") pod "50b63435-18c4-4fcb-821c-3d88abc7b728" (UID: "50b63435-18c4-4fcb-821c-3d88abc7b728"). InnerVolumeSpecName "kube-api-access-4sjtb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:01:22 crc kubenswrapper[4856]: I0126 17:01:22.365042 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50b63435-18c4-4fcb-821c-3d88abc7b728-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "50b63435-18c4-4fcb-821c-3d88abc7b728" (UID: "50b63435-18c4-4fcb-821c-3d88abc7b728"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:01:22 crc kubenswrapper[4856]: I0126 17:01:22.450707 4856 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/50b63435-18c4-4fcb-821c-3d88abc7b728-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 26 17:01:22 crc kubenswrapper[4856]: I0126 17:01:22.450745 4856 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/50b63435-18c4-4fcb-821c-3d88abc7b728-config\") on node \"crc\" DevicePath \"\"" Jan 26 17:01:22 crc kubenswrapper[4856]: I0126 17:01:22.450755 4856 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50b63435-18c4-4fcb-821c-3d88abc7b728-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 17:01:22 crc kubenswrapper[4856]: I0126 17:01:22.450765 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4sjtb\" (UniqueName: \"kubernetes.io/projected/50b63435-18c4-4fcb-821c-3d88abc7b728-kube-api-access-4sjtb\") on node \"crc\" DevicePath \"\"" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.060804 4856 patch_prober.go:28] interesting pod/router-default-5444994796-h9b2g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 17:01:23 crc kubenswrapper[4856]: [-]has-synced failed: reason withheld Jan 26 17:01:23 crc kubenswrapper[4856]: [+]process-running ok Jan 26 17:01:23 crc kubenswrapper[4856]: healthz check failed Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.060868 4856 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9b2g" podUID="85f05bd5-ff83-4d29-9531-ab3499088095" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.227033 4856 generic.go:334] "Generic (PLEG): container finished" podID="9c71e219-35d7-4e1e-a371-3456dfd29e83" containerID="e6e9fc1c7474ee1cf14a50a96e79036f97d946d338d8a18c3434197cbd0438a8" exitCode=0 Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.227161 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mqxwf" event={"ID":"9c71e219-35d7-4e1e-a371-3456dfd29e83","Type":"ContainerDied","Data":"e6e9fc1c7474ee1cf14a50a96e79036f97d946d338d8a18c3434197cbd0438a8"} Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.234868 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-cb449784d-bqprm" event={"ID":"54dee8cd-259a-4c9f-9e56-fbd0ea167f46","Type":"ContainerStarted","Data":"3157d6dd787fe30eefd4db24c3f3619f52444746080c65c699eab5d0d02ab52a"} Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.234926 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-cb449784d-bqprm" event={"ID":"54dee8cd-259a-4c9f-9e56-fbd0ea167f46","Type":"ContainerStarted","Data":"960ff00bedf28636eb04c4f352e2d6d2e33a5ceb9800e901e018103cd5ac5859"} Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.235884 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-cb449784d-bqprm" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.249988 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5457c9f8cf-lcqcl" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.261918 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-cb449784d-bqprm" podStartSLOduration=4.261887795 podStartE2EDuration="4.261887795s" podCreationTimestamp="2026-01-26 17:01:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:01:23.259125884 +0000 UTC m=+179.212379885" watchObservedRunningTime="2026-01-26 17:01:23.261887795 +0000 UTC m=+179.215141776" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.320994 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-cb449784d-bqprm" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.322367 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-54b988dd69-ljwqg"] Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.328928 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5457c9f8cf-lcqcl"] Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.329035 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-54b988dd69-ljwqg" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.334045 4856 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-5457c9f8cf-lcqcl"] Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.334599 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.334785 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.335327 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.336912 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.337547 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.337724 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.339717 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-54b988dd69-ljwqg"] Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.343272 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.439307 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1afc0f4c-e02d-4a70-aaba-e761e8c04eee" path="/var/lib/kubelet/pods/1afc0f4c-e02d-4a70-aaba-e761e8c04eee/volumes" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.440352 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50b63435-18c4-4fcb-821c-3d88abc7b728" path="/var/lib/kubelet/pods/50b63435-18c4-4fcb-821c-3d88abc7b728/volumes" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.443123 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe6baed-ab97-4d8a-8be2-6f00f9698136" path="/var/lib/kubelet/pods/5fe6baed-ab97-4d8a-8be2-6f00f9698136/volumes" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.476025 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/de17aec3-fab1-4a5e-bd46-6a1545b93a89-client-ca\") pod \"controller-manager-54b988dd69-ljwqg\" (UID: \"de17aec3-fab1-4a5e-bd46-6a1545b93a89\") " pod="openshift-controller-manager/controller-manager-54b988dd69-ljwqg" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.476077 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/de17aec3-fab1-4a5e-bd46-6a1545b93a89-proxy-ca-bundles\") pod \"controller-manager-54b988dd69-ljwqg\" (UID: \"de17aec3-fab1-4a5e-bd46-6a1545b93a89\") " pod="openshift-controller-manager/controller-manager-54b988dd69-ljwqg" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.476108 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de17aec3-fab1-4a5e-bd46-6a1545b93a89-config\") pod \"controller-manager-54b988dd69-ljwqg\" (UID: \"de17aec3-fab1-4a5e-bd46-6a1545b93a89\") " pod="openshift-controller-manager/controller-manager-54b988dd69-ljwqg" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.476146 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/de17aec3-fab1-4a5e-bd46-6a1545b93a89-serving-cert\") pod \"controller-manager-54b988dd69-ljwqg\" (UID: \"de17aec3-fab1-4a5e-bd46-6a1545b93a89\") " pod="openshift-controller-manager/controller-manager-54b988dd69-ljwqg" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.476173 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxzpz\" (UniqueName: \"kubernetes.io/projected/de17aec3-fab1-4a5e-bd46-6a1545b93a89-kube-api-access-sxzpz\") pod \"controller-manager-54b988dd69-ljwqg\" (UID: \"de17aec3-fab1-4a5e-bd46-6a1545b93a89\") " pod="openshift-controller-manager/controller-manager-54b988dd69-ljwqg" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.577458 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/de17aec3-fab1-4a5e-bd46-6a1545b93a89-serving-cert\") pod \"controller-manager-54b988dd69-ljwqg\" (UID: \"de17aec3-fab1-4a5e-bd46-6a1545b93a89\") " pod="openshift-controller-manager/controller-manager-54b988dd69-ljwqg" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.577517 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxzpz\" (UniqueName: \"kubernetes.io/projected/de17aec3-fab1-4a5e-bd46-6a1545b93a89-kube-api-access-sxzpz\") pod \"controller-manager-54b988dd69-ljwqg\" (UID: \"de17aec3-fab1-4a5e-bd46-6a1545b93a89\") " pod="openshift-controller-manager/controller-manager-54b988dd69-ljwqg" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.577581 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/de17aec3-fab1-4a5e-bd46-6a1545b93a89-client-ca\") pod \"controller-manager-54b988dd69-ljwqg\" (UID: \"de17aec3-fab1-4a5e-bd46-6a1545b93a89\") " pod="openshift-controller-manager/controller-manager-54b988dd69-ljwqg" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.577604 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/de17aec3-fab1-4a5e-bd46-6a1545b93a89-proxy-ca-bundles\") pod \"controller-manager-54b988dd69-ljwqg\" (UID: \"de17aec3-fab1-4a5e-bd46-6a1545b93a89\") " pod="openshift-controller-manager/controller-manager-54b988dd69-ljwqg" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.577634 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de17aec3-fab1-4a5e-bd46-6a1545b93a89-config\") pod \"controller-manager-54b988dd69-ljwqg\" (UID: \"de17aec3-fab1-4a5e-bd46-6a1545b93a89\") " pod="openshift-controller-manager/controller-manager-54b988dd69-ljwqg" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.578848 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de17aec3-fab1-4a5e-bd46-6a1545b93a89-config\") pod \"controller-manager-54b988dd69-ljwqg\" (UID: \"de17aec3-fab1-4a5e-bd46-6a1545b93a89\") " pod="openshift-controller-manager/controller-manager-54b988dd69-ljwqg" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.580394 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/de17aec3-fab1-4a5e-bd46-6a1545b93a89-client-ca\") pod \"controller-manager-54b988dd69-ljwqg\" (UID: \"de17aec3-fab1-4a5e-bd46-6a1545b93a89\") " pod="openshift-controller-manager/controller-manager-54b988dd69-ljwqg" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.580699 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/de17aec3-fab1-4a5e-bd46-6a1545b93a89-proxy-ca-bundles\") pod \"controller-manager-54b988dd69-ljwqg\" (UID: \"de17aec3-fab1-4a5e-bd46-6a1545b93a89\") " pod="openshift-controller-manager/controller-manager-54b988dd69-ljwqg" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.582984 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/de17aec3-fab1-4a5e-bd46-6a1545b93a89-serving-cert\") pod \"controller-manager-54b988dd69-ljwqg\" (UID: \"de17aec3-fab1-4a5e-bd46-6a1545b93a89\") " pod="openshift-controller-manager/controller-manager-54b988dd69-ljwqg" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.597885 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxzpz\" (UniqueName: \"kubernetes.io/projected/de17aec3-fab1-4a5e-bd46-6a1545b93a89-kube-api-access-sxzpz\") pod \"controller-manager-54b988dd69-ljwqg\" (UID: \"de17aec3-fab1-4a5e-bd46-6a1545b93a89\") " pod="openshift-controller-manager/controller-manager-54b988dd69-ljwqg" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.652173 4856 patch_prober.go:28] interesting pod/console-f9d7485db-6qgnn container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.26:8443/health\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.652248 4856 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-6qgnn" podUID="b28404ed-2e71-4b3f-9140-35ee89dbc8f2" containerName="console" probeResult="failure" output="Get \"https://10.217.0.26:8443/health\": dial tcp 10.217.0.26:8443: connect: connection refused" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.669434 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-54b988dd69-ljwqg" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.694471 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-8q6q4" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.704267 4856 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.709828 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-6rlxp" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.852843 4856 patch_prober.go:28] interesting pod/downloads-7954f5f757-7l927 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.852876 4856 patch_prober.go:28] interesting pod/downloads-7954f5f757-7l927 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.852925 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-7l927" podUID="94291fa4-24a5-499e-8143-89c8784d9284" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.852925 4856 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-7l927" podUID="94291fa4-24a5-499e-8143-89c8784d9284" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.852995 4856 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console/downloads-7954f5f757-7l927" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.853570 4856 patch_prober.go:28] interesting pod/downloads-7954f5f757-7l927 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.853595 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-7l927" podUID="94291fa4-24a5-499e-8143-89c8784d9284" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.853677 4856 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="download-server" containerStatusID={"Type":"cri-o","ID":"d13c7142b05ed798c0e5b16508a221e2918021dbec60054995ac94f05ffdad09"} pod="openshift-console/downloads-7954f5f757-7l927" containerMessage="Container download-server failed liveness probe, will be restarted" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.853721 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/downloads-7954f5f757-7l927" podUID="94291fa4-24a5-499e-8143-89c8784d9284" containerName="download-server" containerID="cri-o://d13c7142b05ed798c0e5b16508a221e2918021dbec60054995ac94f05ffdad09" gracePeriod=2 Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.884953 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kxgfj\" (UniqueName: \"kubernetes.io/projected/7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7-kube-api-access-kxgfj\") pod \"7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7\" (UID: \"7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7\") " Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.885019 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7-config-volume\") pod \"7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7\" (UID: \"7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7\") " Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.885080 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7-secret-volume\") pod \"7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7\" (UID: \"7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7\") " Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.890129 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7-config-volume" (OuterVolumeSpecName: "config-volume") pod "7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7" (UID: "7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.893738 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.895650 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7-kube-api-access-kxgfj" (OuterVolumeSpecName: "kube-api-access-kxgfj") pod "7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7" (UID: "7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7"). InnerVolumeSpecName "kube-api-access-kxgfj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.905246 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-4pbj2" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.906117 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7" (UID: "7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.987066 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kxgfj\" (UniqueName: \"kubernetes.io/projected/7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7-kube-api-access-kxgfj\") on node \"crc\" DevicePath \"\"" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.987096 4856 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7-config-volume\") on node \"crc\" DevicePath \"\"" Jan 26 17:01:23 crc kubenswrapper[4856]: I0126 17:01:23.987105 4856 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 26 17:01:24 crc kubenswrapper[4856]: I0126 17:01:24.069437 4856 patch_prober.go:28] interesting pod/router-default-5444994796-h9b2g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 17:01:24 crc kubenswrapper[4856]: [-]has-synced failed: reason withheld Jan 26 17:01:24 crc kubenswrapper[4856]: [+]process-running ok Jan 26 17:01:24 crc kubenswrapper[4856]: healthz check failed Jan 26 17:01:24 crc kubenswrapper[4856]: I0126 17:01:24.069496 4856 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9b2g" podUID="85f05bd5-ff83-4d29-9531-ab3499088095" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 17:01:24 crc kubenswrapper[4856]: I0126 17:01:24.181849 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-wvttb" Jan 26 17:01:24 crc kubenswrapper[4856]: I0126 17:01:24.250684 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-54b988dd69-ljwqg"] Jan 26 17:01:24 crc kubenswrapper[4856]: I0126 17:01:24.328697 4856 generic.go:334] "Generic (PLEG): container finished" podID="94291fa4-24a5-499e-8143-89c8784d9284" containerID="d13c7142b05ed798c0e5b16508a221e2918021dbec60054995ac94f05ffdad09" exitCode=0 Jan 26 17:01:24 crc kubenswrapper[4856]: I0126 17:01:24.329055 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-7l927" event={"ID":"94291fa4-24a5-499e-8143-89c8784d9284","Type":"ContainerDied","Data":"d13c7142b05ed798c0e5b16508a221e2918021dbec60054995ac94f05ffdad09"} Jan 26 17:01:24 crc kubenswrapper[4856]: I0126 17:01:24.337029 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-8q6q4" Jan 26 17:01:24 crc kubenswrapper[4856]: I0126 17:01:24.345985 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490780-8q6q4" event={"ID":"7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7","Type":"ContainerDied","Data":"d8fe561f33f411cab54065acf50663e1fea5f5209ab612f88976297cc920acef"} Jan 26 17:01:24 crc kubenswrapper[4856]: I0126 17:01:24.346127 4856 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d8fe561f33f411cab54065acf50663e1fea5f5209ab612f88976297cc920acef" Jan 26 17:01:25 crc kubenswrapper[4856]: I0126 17:01:25.060044 4856 patch_prober.go:28] interesting pod/router-default-5444994796-h9b2g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 17:01:25 crc kubenswrapper[4856]: [-]has-synced failed: reason withheld Jan 26 17:01:25 crc kubenswrapper[4856]: [+]process-running ok Jan 26 17:01:25 crc kubenswrapper[4856]: healthz check failed Jan 26 17:01:25 crc kubenswrapper[4856]: I0126 17:01:25.060156 4856 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9b2g" podUID="85f05bd5-ff83-4d29-9531-ab3499088095" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 17:01:25 crc kubenswrapper[4856]: I0126 17:01:25.426786 4856 patch_prober.go:28] interesting pod/downloads-7954f5f757-7l927 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Jan 26 17:01:25 crc kubenswrapper[4856]: I0126 17:01:25.427077 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-7l927" podUID="94291fa4-24a5-499e-8143-89c8784d9284" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Jan 26 17:01:25 crc kubenswrapper[4856]: I0126 17:01:25.447500 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-7l927" event={"ID":"94291fa4-24a5-499e-8143-89c8784d9284","Type":"ContainerStarted","Data":"1a3c8f728acaa63fa83450974ecec3e1e03ae7c892f5036f7c7f018fe224588c"} Jan 26 17:01:25 crc kubenswrapper[4856]: I0126 17:01:25.447591 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-7l927" Jan 26 17:01:25 crc kubenswrapper[4856]: I0126 17:01:25.463548 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-54b988dd69-ljwqg" event={"ID":"de17aec3-fab1-4a5e-bd46-6a1545b93a89","Type":"ContainerStarted","Data":"5dd5f652c6d735efc7f4d83a862afc4f09c88afb0921e6ecb304af76698cc9c8"} Jan 26 17:01:25 crc kubenswrapper[4856]: I0126 17:01:25.463591 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-54b988dd69-ljwqg" event={"ID":"de17aec3-fab1-4a5e-bd46-6a1545b93a89","Type":"ContainerStarted","Data":"1bccd3720f328e7d0b92fc36bcc35726a97ebe6a8070f5cbb1608de57071e2d0"} Jan 26 17:01:25 crc kubenswrapper[4856]: I0126 17:01:25.463610 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-54b988dd69-ljwqg" Jan 26 17:01:25 crc kubenswrapper[4856]: I0126 17:01:25.472359 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-54b988dd69-ljwqg" Jan 26 17:01:26 crc kubenswrapper[4856]: I0126 17:01:26.059974 4856 patch_prober.go:28] interesting pod/router-default-5444994796-h9b2g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 17:01:26 crc kubenswrapper[4856]: [-]has-synced failed: reason withheld Jan 26 17:01:26 crc kubenswrapper[4856]: [+]process-running ok Jan 26 17:01:26 crc kubenswrapper[4856]: healthz check failed Jan 26 17:01:26 crc kubenswrapper[4856]: I0126 17:01:26.060360 4856 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9b2g" podUID="85f05bd5-ff83-4d29-9531-ab3499088095" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 17:01:26 crc kubenswrapper[4856]: I0126 17:01:26.471435 4856 patch_prober.go:28] interesting pod/downloads-7954f5f757-7l927 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Jan 26 17:01:26 crc kubenswrapper[4856]: I0126 17:01:26.471498 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-7l927" podUID="94291fa4-24a5-499e-8143-89c8784d9284" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Jan 26 17:01:26 crc kubenswrapper[4856]: I0126 17:01:26.938413 4856 patch_prober.go:28] interesting pod/machine-config-daemon-xm9cq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:01:26 crc kubenswrapper[4856]: I0126 17:01:26.938478 4856 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:01:27 crc kubenswrapper[4856]: I0126 17:01:27.058701 4856 patch_prober.go:28] interesting pod/router-default-5444994796-h9b2g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 17:01:27 crc kubenswrapper[4856]: [-]has-synced failed: reason withheld Jan 26 17:01:27 crc kubenswrapper[4856]: [+]process-running ok Jan 26 17:01:27 crc kubenswrapper[4856]: healthz check failed Jan 26 17:01:27 crc kubenswrapper[4856]: I0126 17:01:27.058760 4856 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9b2g" podUID="85f05bd5-ff83-4d29-9531-ab3499088095" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 17:01:28 crc kubenswrapper[4856]: I0126 17:01:28.064226 4856 patch_prober.go:28] interesting pod/router-default-5444994796-h9b2g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 17:01:28 crc kubenswrapper[4856]: [-]has-synced failed: reason withheld Jan 26 17:01:28 crc kubenswrapper[4856]: [+]process-running ok Jan 26 17:01:28 crc kubenswrapper[4856]: healthz check failed Jan 26 17:01:28 crc kubenswrapper[4856]: I0126 17:01:28.064874 4856 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9b2g" podUID="85f05bd5-ff83-4d29-9531-ab3499088095" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 17:01:29 crc kubenswrapper[4856]: I0126 17:01:29.562093 4856 patch_prober.go:28] interesting pod/router-default-5444994796-h9b2g container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 17:01:29 crc kubenswrapper[4856]: [+]has-synced ok Jan 26 17:01:29 crc kubenswrapper[4856]: [+]process-running ok Jan 26 17:01:29 crc kubenswrapper[4856]: healthz check failed Jan 26 17:01:29 crc kubenswrapper[4856]: I0126 17:01:29.562177 4856 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-h9b2g" podUID="85f05bd5-ff83-4d29-9531-ab3499088095" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 17:01:30 crc kubenswrapper[4856]: I0126 17:01:30.057284 4856 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-h9b2g" Jan 26 17:01:30 crc kubenswrapper[4856]: I0126 17:01:30.060416 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-h9b2g" Jan 26 17:01:30 crc kubenswrapper[4856]: I0126 17:01:30.082325 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-54b988dd69-ljwqg" podStartSLOduration=9.082308566 podStartE2EDuration="9.082308566s" podCreationTimestamp="2026-01-26 17:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:01:25.534987137 +0000 UTC m=+181.488241118" watchObservedRunningTime="2026-01-26 17:01:30.082308566 +0000 UTC m=+186.035562537" Jan 26 17:01:33 crc kubenswrapper[4856]: I0126 17:01:33.858941 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-7l927" Jan 26 17:01:34 crc kubenswrapper[4856]: I0126 17:01:34.152941 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-8m4l6" Jan 26 17:01:35 crc kubenswrapper[4856]: I0126 17:01:35.095759 4856 patch_prober.go:28] interesting pod/router-default-5444994796-h9b2g container/router namespace/openshift-ingress: Readiness probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 26 17:01:35 crc kubenswrapper[4856]: I0126 17:01:35.095929 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5444994796-h9b2g" podUID="85f05bd5-ff83-4d29-9531-ab3499088095" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 26 17:01:35 crc kubenswrapper[4856]: I0126 17:01:35.109704 4856 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-6qgnn" Jan 26 17:01:35 crc kubenswrapper[4856]: I0126 17:01:35.114211 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-6qgnn" Jan 26 17:01:39 crc kubenswrapper[4856]: I0126 17:01:39.428604 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:01:43 crc kubenswrapper[4856]: I0126 17:01:43.007808 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 17:01:54 crc kubenswrapper[4856]: I0126 17:01:54.059150 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 26 17:01:54 crc kubenswrapper[4856]: E0126 17:01:54.059877 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7" containerName="collect-profiles" Jan 26 17:01:54 crc kubenswrapper[4856]: I0126 17:01:54.059890 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7" containerName="collect-profiles" Jan 26 17:01:54 crc kubenswrapper[4856]: I0126 17:01:54.059996 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e9ee376-b7c7-4b6a-91b3-cc86a3a02dc7" containerName="collect-profiles" Jan 26 17:01:54 crc kubenswrapper[4856]: I0126 17:01:54.060355 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 26 17:01:54 crc kubenswrapper[4856]: I0126 17:01:54.067915 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Jan 26 17:01:54 crc kubenswrapper[4856]: I0126 17:01:54.068938 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Jan 26 17:01:54 crc kubenswrapper[4856]: I0126 17:01:54.072874 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 26 17:01:54 crc kubenswrapper[4856]: I0126 17:01:54.073175 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/55a1283a-85e0-497f-8c5d-9a28168cb810-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"55a1283a-85e0-497f-8c5d-9a28168cb810\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 26 17:01:54 crc kubenswrapper[4856]: I0126 17:01:54.073239 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/55a1283a-85e0-497f-8c5d-9a28168cb810-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"55a1283a-85e0-497f-8c5d-9a28168cb810\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 26 17:01:54 crc kubenswrapper[4856]: I0126 17:01:54.174039 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/55a1283a-85e0-497f-8c5d-9a28168cb810-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"55a1283a-85e0-497f-8c5d-9a28168cb810\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 26 17:01:54 crc kubenswrapper[4856]: I0126 17:01:54.174129 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/55a1283a-85e0-497f-8c5d-9a28168cb810-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"55a1283a-85e0-497f-8c5d-9a28168cb810\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 26 17:01:54 crc kubenswrapper[4856]: I0126 17:01:54.174210 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/55a1283a-85e0-497f-8c5d-9a28168cb810-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"55a1283a-85e0-497f-8c5d-9a28168cb810\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 26 17:01:54 crc kubenswrapper[4856]: I0126 17:01:54.196116 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/55a1283a-85e0-497f-8c5d-9a28168cb810-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"55a1283a-85e0-497f-8c5d-9a28168cb810\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 26 17:01:54 crc kubenswrapper[4856]: I0126 17:01:54.388582 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 26 17:01:56 crc kubenswrapper[4856]: I0126 17:01:56.939109 4856 patch_prober.go:28] interesting pod/machine-config-daemon-xm9cq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:01:56 crc kubenswrapper[4856]: I0126 17:01:56.939435 4856 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:01:58 crc kubenswrapper[4856]: I0126 17:01:58.461873 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 26 17:01:58 crc kubenswrapper[4856]: I0126 17:01:58.466901 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 26 17:01:58 crc kubenswrapper[4856]: I0126 17:01:58.471180 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 26 17:01:58 crc kubenswrapper[4856]: I0126 17:01:58.532670 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/69379820-3062-4964-a8dd-8689f8cea38d-kube-api-access\") pod \"installer-9-crc\" (UID: \"69379820-3062-4964-a8dd-8689f8cea38d\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 26 17:01:58 crc kubenswrapper[4856]: I0126 17:01:58.532727 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/69379820-3062-4964-a8dd-8689f8cea38d-var-lock\") pod \"installer-9-crc\" (UID: \"69379820-3062-4964-a8dd-8689f8cea38d\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 26 17:01:58 crc kubenswrapper[4856]: I0126 17:01:58.532792 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/69379820-3062-4964-a8dd-8689f8cea38d-kubelet-dir\") pod \"installer-9-crc\" (UID: \"69379820-3062-4964-a8dd-8689f8cea38d\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 26 17:01:58 crc kubenswrapper[4856]: I0126 17:01:58.634041 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/69379820-3062-4964-a8dd-8689f8cea38d-var-lock\") pod \"installer-9-crc\" (UID: \"69379820-3062-4964-a8dd-8689f8cea38d\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 26 17:01:58 crc kubenswrapper[4856]: I0126 17:01:58.634146 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/69379820-3062-4964-a8dd-8689f8cea38d-kubelet-dir\") pod \"installer-9-crc\" (UID: \"69379820-3062-4964-a8dd-8689f8cea38d\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 26 17:01:58 crc kubenswrapper[4856]: I0126 17:01:58.634201 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/69379820-3062-4964-a8dd-8689f8cea38d-var-lock\") pod \"installer-9-crc\" (UID: \"69379820-3062-4964-a8dd-8689f8cea38d\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 26 17:01:58 crc kubenswrapper[4856]: I0126 17:01:58.634237 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/69379820-3062-4964-a8dd-8689f8cea38d-kube-api-access\") pod \"installer-9-crc\" (UID: \"69379820-3062-4964-a8dd-8689f8cea38d\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 26 17:01:58 crc kubenswrapper[4856]: I0126 17:01:58.634366 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/69379820-3062-4964-a8dd-8689f8cea38d-kubelet-dir\") pod \"installer-9-crc\" (UID: \"69379820-3062-4964-a8dd-8689f8cea38d\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 26 17:01:58 crc kubenswrapper[4856]: I0126 17:01:58.654018 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/69379820-3062-4964-a8dd-8689f8cea38d-kube-api-access\") pod \"installer-9-crc\" (UID: \"69379820-3062-4964-a8dd-8689f8cea38d\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 26 17:01:58 crc kubenswrapper[4856]: I0126 17:01:58.802979 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 26 17:02:15 crc kubenswrapper[4856]: E0126 17:02:15.160269 4856 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 26 17:02:15 crc kubenswrapper[4856]: E0126 17:02:15.161834 4856 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tf49w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-txmdl_openshift-marketplace(40a27476-22b1-4083-990e-66e70ccdaf4c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 26 17:02:15 crc kubenswrapper[4856]: E0126 17:02:15.163091 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-txmdl" podUID="40a27476-22b1-4083-990e-66e70ccdaf4c" Jan 26 17:02:16 crc kubenswrapper[4856]: E0126 17:02:16.589664 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-txmdl" podUID="40a27476-22b1-4083-990e-66e70ccdaf4c" Jan 26 17:02:16 crc kubenswrapper[4856]: E0126 17:02:16.658808 4856 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 26 17:02:16 crc kubenswrapper[4856]: E0126 17:02:16.659000 4856 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2qzgz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-4kwt4_openshift-marketplace(d6944fc9-b8d7-4013-8702-b5765c410a0b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 26 17:02:16 crc kubenswrapper[4856]: E0126 17:02:16.660254 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-4kwt4" podUID="d6944fc9-b8d7-4013-8702-b5765c410a0b" Jan 26 17:02:16 crc kubenswrapper[4856]: E0126 17:02:16.687771 4856 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Jan 26 17:02:16 crc kubenswrapper[4856]: E0126 17:02:16.688070 4856 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s92lp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-62nhd_openshift-marketplace(7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 26 17:02:16 crc kubenswrapper[4856]: E0126 17:02:16.689273 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-62nhd" podUID="7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766" Jan 26 17:02:20 crc kubenswrapper[4856]: E0126 17:02:20.666250 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-4kwt4" podUID="d6944fc9-b8d7-4013-8702-b5765c410a0b" Jan 26 17:02:20 crc kubenswrapper[4856]: E0126 17:02:20.666308 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-62nhd" podUID="7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766" Jan 26 17:02:20 crc kubenswrapper[4856]: E0126 17:02:20.769112 4856 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 26 17:02:20 crc kubenswrapper[4856]: E0126 17:02:20.769303 4856 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wvpzb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-qknj9_openshift-marketplace(a3fa94fe-e4ad-4171-b853-89878dc61569): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 26 17:02:20 crc kubenswrapper[4856]: E0126 17:02:20.770485 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-qknj9" podUID="a3fa94fe-e4ad-4171-b853-89878dc61569" Jan 26 17:02:22 crc kubenswrapper[4856]: E0126 17:02:22.116363 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-qknj9" podUID="a3fa94fe-e4ad-4171-b853-89878dc61569" Jan 26 17:02:22 crc kubenswrapper[4856]: E0126 17:02:22.190746 4856 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 26 17:02:22 crc kubenswrapper[4856]: E0126 17:02:22.190909 4856 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x7mvx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-n8hp2_openshift-marketplace(a6086d4b-faeb-4a12-8e6a-2a178dfe374c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 26 17:02:22 crc kubenswrapper[4856]: E0126 17:02:22.192096 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-n8hp2" podUID="a6086d4b-faeb-4a12-8e6a-2a178dfe374c" Jan 26 17:02:22 crc kubenswrapper[4856]: E0126 17:02:22.203986 4856 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 26 17:02:22 crc kubenswrapper[4856]: E0126 17:02:22.204442 4856 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dcbvc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-g8bgt_openshift-marketplace(0d7eb7b8-63ae-493a-850b-0b9f3b42e927): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 26 17:02:22 crc kubenswrapper[4856]: E0126 17:02:22.205610 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-g8bgt" podUID="0d7eb7b8-63ae-493a-850b-0b9f3b42e927" Jan 26 17:02:22 crc kubenswrapper[4856]: E0126 17:02:22.234717 4856 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Jan 26 17:02:22 crc kubenswrapper[4856]: E0126 17:02:22.234931 4856 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6swzw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-mqxwf_openshift-marketplace(9c71e219-35d7-4e1e-a371-3456dfd29e83): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 26 17:02:22 crc kubenswrapper[4856]: E0126 17:02:22.239436 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-mqxwf" podUID="9c71e219-35d7-4e1e-a371-3456dfd29e83" Jan 26 17:02:22 crc kubenswrapper[4856]: E0126 17:02:22.252702 4856 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 26 17:02:22 crc kubenswrapper[4856]: E0126 17:02:22.253142 4856 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gtxjz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-qgjjd_openshift-marketplace(89cf05de-642b-4574-9f79-45e7a3d4afa3): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 26 17:02:22 crc kubenswrapper[4856]: E0126 17:02:22.254322 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-qgjjd" podUID="89cf05de-642b-4574-9f79-45e7a3d4afa3" Jan 26 17:02:22 crc kubenswrapper[4856]: I0126 17:02:22.528874 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 26 17:02:22 crc kubenswrapper[4856]: I0126 17:02:22.592163 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 26 17:02:22 crc kubenswrapper[4856]: W0126 17:02:22.607358 4856 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod69379820_3062_4964_a8dd_8689f8cea38d.slice/crio-af60ab5d4a2b57ad1bbcc4a879fdc9dce5f1b3ef1e2f5eb96e13241cdf6f2277 WatchSource:0}: Error finding container af60ab5d4a2b57ad1bbcc4a879fdc9dce5f1b3ef1e2f5eb96e13241cdf6f2277: Status 404 returned error can't find the container with id af60ab5d4a2b57ad1bbcc4a879fdc9dce5f1b3ef1e2f5eb96e13241cdf6f2277 Jan 26 17:02:22 crc kubenswrapper[4856]: I0126 17:02:22.948031 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"55a1283a-85e0-497f-8c5d-9a28168cb810","Type":"ContainerStarted","Data":"7d4c917e0830be1aeb1e35a23c1be0bcb5487e9bceffc19d68bcd09d48f247ad"} Jan 26 17:02:22 crc kubenswrapper[4856]: I0126 17:02:22.948400 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"55a1283a-85e0-497f-8c5d-9a28168cb810","Type":"ContainerStarted","Data":"c8b87a1aff72cc4fdfed70fd560545e58b33f1c048b47cb99c85517d64eba518"} Jan 26 17:02:22 crc kubenswrapper[4856]: I0126 17:02:22.949898 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"69379820-3062-4964-a8dd-8689f8cea38d","Type":"ContainerStarted","Data":"beeb8e8929ad597a53e5bcbe203dbd0aeea7fb6f4cfbcd350384cfbddded9459"} Jan 26 17:02:22 crc kubenswrapper[4856]: I0126 17:02:22.949956 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"69379820-3062-4964-a8dd-8689f8cea38d","Type":"ContainerStarted","Data":"af60ab5d4a2b57ad1bbcc4a879fdc9dce5f1b3ef1e2f5eb96e13241cdf6f2277"} Jan 26 17:02:22 crc kubenswrapper[4856]: E0126 17:02:22.953405 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-qgjjd" podUID="89cf05de-642b-4574-9f79-45e7a3d4afa3" Jan 26 17:02:22 crc kubenswrapper[4856]: E0126 17:02:22.955970 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-n8hp2" podUID="a6086d4b-faeb-4a12-8e6a-2a178dfe374c" Jan 26 17:02:22 crc kubenswrapper[4856]: E0126 17:02:22.958003 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-g8bgt" podUID="0d7eb7b8-63ae-493a-850b-0b9f3b42e927" Jan 26 17:02:22 crc kubenswrapper[4856]: E0126 17:02:22.969756 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-mqxwf" podUID="9c71e219-35d7-4e1e-a371-3456dfd29e83" Jan 26 17:02:23 crc kubenswrapper[4856]: I0126 17:02:23.958132 4856 generic.go:334] "Generic (PLEG): container finished" podID="55a1283a-85e0-497f-8c5d-9a28168cb810" containerID="7d4c917e0830be1aeb1e35a23c1be0bcb5487e9bceffc19d68bcd09d48f247ad" exitCode=0 Jan 26 17:02:23 crc kubenswrapper[4856]: I0126 17:02:23.958201 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"55a1283a-85e0-497f-8c5d-9a28168cb810","Type":"ContainerDied","Data":"7d4c917e0830be1aeb1e35a23c1be0bcb5487e9bceffc19d68bcd09d48f247ad"} Jan 26 17:02:23 crc kubenswrapper[4856]: I0126 17:02:23.991611 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=25.991585429 podStartE2EDuration="25.991585429s" podCreationTimestamp="2026-01-26 17:01:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:02:23.989468278 +0000 UTC m=+239.942722299" watchObservedRunningTime="2026-01-26 17:02:23.991585429 +0000 UTC m=+239.944839420" Jan 26 17:02:25 crc kubenswrapper[4856]: I0126 17:02:25.221395 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 26 17:02:25 crc kubenswrapper[4856]: I0126 17:02:25.405909 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/55a1283a-85e0-497f-8c5d-9a28168cb810-kube-api-access\") pod \"55a1283a-85e0-497f-8c5d-9a28168cb810\" (UID: \"55a1283a-85e0-497f-8c5d-9a28168cb810\") " Jan 26 17:02:25 crc kubenswrapper[4856]: I0126 17:02:25.406030 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/55a1283a-85e0-497f-8c5d-9a28168cb810-kubelet-dir\") pod \"55a1283a-85e0-497f-8c5d-9a28168cb810\" (UID: \"55a1283a-85e0-497f-8c5d-9a28168cb810\") " Jan 26 17:02:25 crc kubenswrapper[4856]: I0126 17:02:25.406073 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/55a1283a-85e0-497f-8c5d-9a28168cb810-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "55a1283a-85e0-497f-8c5d-9a28168cb810" (UID: "55a1283a-85e0-497f-8c5d-9a28168cb810"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:02:25 crc kubenswrapper[4856]: I0126 17:02:25.406653 4856 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/55a1283a-85e0-497f-8c5d-9a28168cb810-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 26 17:02:25 crc kubenswrapper[4856]: I0126 17:02:25.411692 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55a1283a-85e0-497f-8c5d-9a28168cb810-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "55a1283a-85e0-497f-8c5d-9a28168cb810" (UID: "55a1283a-85e0-497f-8c5d-9a28168cb810"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:02:25 crc kubenswrapper[4856]: I0126 17:02:25.508413 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/55a1283a-85e0-497f-8c5d-9a28168cb810-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 26 17:02:25 crc kubenswrapper[4856]: I0126 17:02:25.975543 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"55a1283a-85e0-497f-8c5d-9a28168cb810","Type":"ContainerDied","Data":"c8b87a1aff72cc4fdfed70fd560545e58b33f1c048b47cb99c85517d64eba518"} Jan 26 17:02:25 crc kubenswrapper[4856]: I0126 17:02:25.975590 4856 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c8b87a1aff72cc4fdfed70fd560545e58b33f1c048b47cb99c85517d64eba518" Jan 26 17:02:25 crc kubenswrapper[4856]: I0126 17:02:25.975671 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 26 17:02:26 crc kubenswrapper[4856]: I0126 17:02:26.939277 4856 patch_prober.go:28] interesting pod/machine-config-daemon-xm9cq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:02:26 crc kubenswrapper[4856]: I0126 17:02:26.939332 4856 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:02:26 crc kubenswrapper[4856]: I0126 17:02:26.939380 4856 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" Jan 26 17:02:26 crc kubenswrapper[4856]: I0126 17:02:26.939910 4856 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"54ca3fa13d9e8d442efa93b44a870369f7df3fe7562d77b98528f5c19a751f18"} pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 26 17:02:26 crc kubenswrapper[4856]: I0126 17:02:26.939954 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" containerName="machine-config-daemon" containerID="cri-o://54ca3fa13d9e8d442efa93b44a870369f7df3fe7562d77b98528f5c19a751f18" gracePeriod=600 Jan 26 17:02:27 crc kubenswrapper[4856]: I0126 17:02:27.987135 4856 generic.go:334] "Generic (PLEG): container finished" podID="63c75ede-5170-4db0-811b-5217ef8d72b3" containerID="54ca3fa13d9e8d442efa93b44a870369f7df3fe7562d77b98528f5c19a751f18" exitCode=0 Jan 26 17:02:27 crc kubenswrapper[4856]: I0126 17:02:27.987301 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" event={"ID":"63c75ede-5170-4db0-811b-5217ef8d72b3","Type":"ContainerDied","Data":"54ca3fa13d9e8d442efa93b44a870369f7df3fe7562d77b98528f5c19a751f18"} Jan 26 17:02:29 crc kubenswrapper[4856]: I0126 17:02:29.538678 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" event={"ID":"63c75ede-5170-4db0-811b-5217ef8d72b3","Type":"ContainerStarted","Data":"9758bfdfd1807e791935ac7ec93246863e5867351e35d27ffaff68ae79110e9c"} Jan 26 17:02:30 crc kubenswrapper[4856]: I0126 17:02:30.537074 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-txmdl" event={"ID":"40a27476-22b1-4083-990e-66e70ccdaf4c","Type":"ContainerStarted","Data":"4be8cc185ffcb38acac0516b9ba74f7fa439552b3ff463ccf813f91341bce48c"} Jan 26 17:02:31 crc kubenswrapper[4856]: I0126 17:02:31.543467 4856 generic.go:334] "Generic (PLEG): container finished" podID="40a27476-22b1-4083-990e-66e70ccdaf4c" containerID="4be8cc185ffcb38acac0516b9ba74f7fa439552b3ff463ccf813f91341bce48c" exitCode=0 Jan 26 17:02:31 crc kubenswrapper[4856]: I0126 17:02:31.543555 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-txmdl" event={"ID":"40a27476-22b1-4083-990e-66e70ccdaf4c","Type":"ContainerDied","Data":"4be8cc185ffcb38acac0516b9ba74f7fa439552b3ff463ccf813f91341bce48c"} Jan 26 17:02:33 crc kubenswrapper[4856]: I0126 17:02:33.556845 4856 generic.go:334] "Generic (PLEG): container finished" podID="d6944fc9-b8d7-4013-8702-b5765c410a0b" containerID="25c95d58185d3429ff473dbc3a21342905624b31dd17e96573eb140be4c2402c" exitCode=0 Jan 26 17:02:33 crc kubenswrapper[4856]: I0126 17:02:33.558214 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4kwt4" event={"ID":"d6944fc9-b8d7-4013-8702-b5765c410a0b","Type":"ContainerDied","Data":"25c95d58185d3429ff473dbc3a21342905624b31dd17e96573eb140be4c2402c"} Jan 26 17:02:33 crc kubenswrapper[4856]: I0126 17:02:33.561404 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-txmdl" event={"ID":"40a27476-22b1-4083-990e-66e70ccdaf4c","Type":"ContainerStarted","Data":"d2eb8e794ba046c4faf481951340700b775e721a354e0ab5bea576c03e396810"} Jan 26 17:02:33 crc kubenswrapper[4856]: I0126 17:02:33.566884 4856 generic.go:334] "Generic (PLEG): container finished" podID="7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766" containerID="d684603f69b61a1ce87ec7d1d3ef00e518372571ee64ede6a51ce75afd2227ca" exitCode=0 Jan 26 17:02:33 crc kubenswrapper[4856]: I0126 17:02:33.566929 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-62nhd" event={"ID":"7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766","Type":"ContainerDied","Data":"d684603f69b61a1ce87ec7d1d3ef00e518372571ee64ede6a51ce75afd2227ca"} Jan 26 17:02:33 crc kubenswrapper[4856]: I0126 17:02:33.782186 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-txmdl" podStartSLOduration=5.360780902 podStartE2EDuration="1m16.782163275s" podCreationTimestamp="2026-01-26 17:01:17 +0000 UTC" firstStartedPulling="2026-01-26 17:01:21.086300628 +0000 UTC m=+177.039554609" lastFinishedPulling="2026-01-26 17:02:32.507683001 +0000 UTC m=+248.460936982" observedRunningTime="2026-01-26 17:02:33.630036358 +0000 UTC m=+249.583290339" watchObservedRunningTime="2026-01-26 17:02:33.782163275 +0000 UTC m=+249.735417256" Jan 26 17:02:33 crc kubenswrapper[4856]: I0126 17:02:33.783682 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-dnvfq"] Jan 26 17:02:33 crc kubenswrapper[4856]: E0126 17:02:33.783984 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55a1283a-85e0-497f-8c5d-9a28168cb810" containerName="pruner" Jan 26 17:02:33 crc kubenswrapper[4856]: I0126 17:02:33.784005 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="55a1283a-85e0-497f-8c5d-9a28168cb810" containerName="pruner" Jan 26 17:02:33 crc kubenswrapper[4856]: I0126 17:02:33.784143 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="55a1283a-85e0-497f-8c5d-9a28168cb810" containerName="pruner" Jan 26 17:02:33 crc kubenswrapper[4856]: I0126 17:02:33.784716 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-dnvfq" Jan 26 17:02:33 crc kubenswrapper[4856]: I0126 17:02:33.805553 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-dnvfq"] Jan 26 17:02:33 crc kubenswrapper[4856]: I0126 17:02:33.886417 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/26fc27be-aaf6-4ce1-9b90-0bca184e8f12-trusted-ca\") pod \"image-registry-66df7c8f76-dnvfq\" (UID: \"26fc27be-aaf6-4ce1-9b90-0bca184e8f12\") " pod="openshift-image-registry/image-registry-66df7c8f76-dnvfq" Jan 26 17:02:33 crc kubenswrapper[4856]: I0126 17:02:33.886748 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/26fc27be-aaf6-4ce1-9b90-0bca184e8f12-registry-certificates\") pod \"image-registry-66df7c8f76-dnvfq\" (UID: \"26fc27be-aaf6-4ce1-9b90-0bca184e8f12\") " pod="openshift-image-registry/image-registry-66df7c8f76-dnvfq" Jan 26 17:02:33 crc kubenswrapper[4856]: I0126 17:02:33.886900 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-dnvfq\" (UID: \"26fc27be-aaf6-4ce1-9b90-0bca184e8f12\") " pod="openshift-image-registry/image-registry-66df7c8f76-dnvfq" Jan 26 17:02:33 crc kubenswrapper[4856]: I0126 17:02:33.887086 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/26fc27be-aaf6-4ce1-9b90-0bca184e8f12-ca-trust-extracted\") pod \"image-registry-66df7c8f76-dnvfq\" (UID: \"26fc27be-aaf6-4ce1-9b90-0bca184e8f12\") " pod="openshift-image-registry/image-registry-66df7c8f76-dnvfq" Jan 26 17:02:33 crc kubenswrapper[4856]: I0126 17:02:33.887264 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/26fc27be-aaf6-4ce1-9b90-0bca184e8f12-registry-tls\") pod \"image-registry-66df7c8f76-dnvfq\" (UID: \"26fc27be-aaf6-4ce1-9b90-0bca184e8f12\") " pod="openshift-image-registry/image-registry-66df7c8f76-dnvfq" Jan 26 17:02:33 crc kubenswrapper[4856]: I0126 17:02:33.887427 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/26fc27be-aaf6-4ce1-9b90-0bca184e8f12-bound-sa-token\") pod \"image-registry-66df7c8f76-dnvfq\" (UID: \"26fc27be-aaf6-4ce1-9b90-0bca184e8f12\") " pod="openshift-image-registry/image-registry-66df7c8f76-dnvfq" Jan 26 17:02:33 crc kubenswrapper[4856]: I0126 17:02:33.887592 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/26fc27be-aaf6-4ce1-9b90-0bca184e8f12-installation-pull-secrets\") pod \"image-registry-66df7c8f76-dnvfq\" (UID: \"26fc27be-aaf6-4ce1-9b90-0bca184e8f12\") " pod="openshift-image-registry/image-registry-66df7c8f76-dnvfq" Jan 26 17:02:33 crc kubenswrapper[4856]: I0126 17:02:33.887744 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmbf7\" (UniqueName: \"kubernetes.io/projected/26fc27be-aaf6-4ce1-9b90-0bca184e8f12-kube-api-access-mmbf7\") pod \"image-registry-66df7c8f76-dnvfq\" (UID: \"26fc27be-aaf6-4ce1-9b90-0bca184e8f12\") " pod="openshift-image-registry/image-registry-66df7c8f76-dnvfq" Jan 26 17:02:33 crc kubenswrapper[4856]: I0126 17:02:33.920661 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-dnvfq\" (UID: \"26fc27be-aaf6-4ce1-9b90-0bca184e8f12\") " pod="openshift-image-registry/image-registry-66df7c8f76-dnvfq" Jan 26 17:02:33 crc kubenswrapper[4856]: I0126 17:02:33.989181 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmbf7\" (UniqueName: \"kubernetes.io/projected/26fc27be-aaf6-4ce1-9b90-0bca184e8f12-kube-api-access-mmbf7\") pod \"image-registry-66df7c8f76-dnvfq\" (UID: \"26fc27be-aaf6-4ce1-9b90-0bca184e8f12\") " pod="openshift-image-registry/image-registry-66df7c8f76-dnvfq" Jan 26 17:02:33 crc kubenswrapper[4856]: I0126 17:02:33.989239 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/26fc27be-aaf6-4ce1-9b90-0bca184e8f12-trusted-ca\") pod \"image-registry-66df7c8f76-dnvfq\" (UID: \"26fc27be-aaf6-4ce1-9b90-0bca184e8f12\") " pod="openshift-image-registry/image-registry-66df7c8f76-dnvfq" Jan 26 17:02:33 crc kubenswrapper[4856]: I0126 17:02:33.989281 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/26fc27be-aaf6-4ce1-9b90-0bca184e8f12-registry-certificates\") pod \"image-registry-66df7c8f76-dnvfq\" (UID: \"26fc27be-aaf6-4ce1-9b90-0bca184e8f12\") " pod="openshift-image-registry/image-registry-66df7c8f76-dnvfq" Jan 26 17:02:33 crc kubenswrapper[4856]: I0126 17:02:33.989341 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/26fc27be-aaf6-4ce1-9b90-0bca184e8f12-ca-trust-extracted\") pod \"image-registry-66df7c8f76-dnvfq\" (UID: \"26fc27be-aaf6-4ce1-9b90-0bca184e8f12\") " pod="openshift-image-registry/image-registry-66df7c8f76-dnvfq" Jan 26 17:02:33 crc kubenswrapper[4856]: I0126 17:02:33.989375 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/26fc27be-aaf6-4ce1-9b90-0bca184e8f12-registry-tls\") pod \"image-registry-66df7c8f76-dnvfq\" (UID: \"26fc27be-aaf6-4ce1-9b90-0bca184e8f12\") " pod="openshift-image-registry/image-registry-66df7c8f76-dnvfq" Jan 26 17:02:33 crc kubenswrapper[4856]: I0126 17:02:33.989411 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/26fc27be-aaf6-4ce1-9b90-0bca184e8f12-bound-sa-token\") pod \"image-registry-66df7c8f76-dnvfq\" (UID: \"26fc27be-aaf6-4ce1-9b90-0bca184e8f12\") " pod="openshift-image-registry/image-registry-66df7c8f76-dnvfq" Jan 26 17:02:33 crc kubenswrapper[4856]: I0126 17:02:33.989437 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/26fc27be-aaf6-4ce1-9b90-0bca184e8f12-installation-pull-secrets\") pod \"image-registry-66df7c8f76-dnvfq\" (UID: \"26fc27be-aaf6-4ce1-9b90-0bca184e8f12\") " pod="openshift-image-registry/image-registry-66df7c8f76-dnvfq" Jan 26 17:02:33 crc kubenswrapper[4856]: I0126 17:02:33.990935 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/26fc27be-aaf6-4ce1-9b90-0bca184e8f12-ca-trust-extracted\") pod \"image-registry-66df7c8f76-dnvfq\" (UID: \"26fc27be-aaf6-4ce1-9b90-0bca184e8f12\") " pod="openshift-image-registry/image-registry-66df7c8f76-dnvfq" Jan 26 17:02:33 crc kubenswrapper[4856]: I0126 17:02:33.991414 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/26fc27be-aaf6-4ce1-9b90-0bca184e8f12-trusted-ca\") pod \"image-registry-66df7c8f76-dnvfq\" (UID: \"26fc27be-aaf6-4ce1-9b90-0bca184e8f12\") " pod="openshift-image-registry/image-registry-66df7c8f76-dnvfq" Jan 26 17:02:33 crc kubenswrapper[4856]: I0126 17:02:33.994032 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/26fc27be-aaf6-4ce1-9b90-0bca184e8f12-registry-certificates\") pod \"image-registry-66df7c8f76-dnvfq\" (UID: \"26fc27be-aaf6-4ce1-9b90-0bca184e8f12\") " pod="openshift-image-registry/image-registry-66df7c8f76-dnvfq" Jan 26 17:02:33 crc kubenswrapper[4856]: I0126 17:02:33.995441 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/26fc27be-aaf6-4ce1-9b90-0bca184e8f12-installation-pull-secrets\") pod \"image-registry-66df7c8f76-dnvfq\" (UID: \"26fc27be-aaf6-4ce1-9b90-0bca184e8f12\") " pod="openshift-image-registry/image-registry-66df7c8f76-dnvfq" Jan 26 17:02:33 crc kubenswrapper[4856]: I0126 17:02:33.995467 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/26fc27be-aaf6-4ce1-9b90-0bca184e8f12-registry-tls\") pod \"image-registry-66df7c8f76-dnvfq\" (UID: \"26fc27be-aaf6-4ce1-9b90-0bca184e8f12\") " pod="openshift-image-registry/image-registry-66df7c8f76-dnvfq" Jan 26 17:02:34 crc kubenswrapper[4856]: I0126 17:02:34.026840 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmbf7\" (UniqueName: \"kubernetes.io/projected/26fc27be-aaf6-4ce1-9b90-0bca184e8f12-kube-api-access-mmbf7\") pod \"image-registry-66df7c8f76-dnvfq\" (UID: \"26fc27be-aaf6-4ce1-9b90-0bca184e8f12\") " pod="openshift-image-registry/image-registry-66df7c8f76-dnvfq" Jan 26 17:02:34 crc kubenswrapper[4856]: I0126 17:02:34.028620 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/26fc27be-aaf6-4ce1-9b90-0bca184e8f12-bound-sa-token\") pod \"image-registry-66df7c8f76-dnvfq\" (UID: \"26fc27be-aaf6-4ce1-9b90-0bca184e8f12\") " pod="openshift-image-registry/image-registry-66df7c8f76-dnvfq" Jan 26 17:02:34 crc kubenswrapper[4856]: I0126 17:02:34.100704 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-dnvfq" Jan 26 17:02:34 crc kubenswrapper[4856]: I0126 17:02:34.521698 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-dnvfq"] Jan 26 17:02:34 crc kubenswrapper[4856]: W0126 17:02:34.524925 4856 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod26fc27be_aaf6_4ce1_9b90_0bca184e8f12.slice/crio-04a92d7e0364d4e4c12781d3d1a2e91386de31dbf935f68165ee75d7491184f3 WatchSource:0}: Error finding container 04a92d7e0364d4e4c12781d3d1a2e91386de31dbf935f68165ee75d7491184f3: Status 404 returned error can't find the container with id 04a92d7e0364d4e4c12781d3d1a2e91386de31dbf935f68165ee75d7491184f3 Jan 26 17:02:34 crc kubenswrapper[4856]: I0126 17:02:34.574335 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-62nhd" event={"ID":"7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766","Type":"ContainerStarted","Data":"67c41d7af13d33af9423d069e86b531ff9d226b1435b62347517f490f3904943"} Jan 26 17:02:34 crc kubenswrapper[4856]: I0126 17:02:34.577512 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4kwt4" event={"ID":"d6944fc9-b8d7-4013-8702-b5765c410a0b","Type":"ContainerStarted","Data":"a6302eb5e39718f049dce88c0f2a26632538c4eb99b7dbea0dca8f7aae8306c7"} Jan 26 17:02:34 crc kubenswrapper[4856]: I0126 17:02:34.579759 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-dnvfq" event={"ID":"26fc27be-aaf6-4ce1-9b90-0bca184e8f12","Type":"ContainerStarted","Data":"04a92d7e0364d4e4c12781d3d1a2e91386de31dbf935f68165ee75d7491184f3"} Jan 26 17:02:34 crc kubenswrapper[4856]: I0126 17:02:34.601297 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-62nhd" podStartSLOduration=4.791239977 podStartE2EDuration="1m17.601280986s" podCreationTimestamp="2026-01-26 17:01:17 +0000 UTC" firstStartedPulling="2026-01-26 17:01:21.14980898 +0000 UTC m=+177.103062961" lastFinishedPulling="2026-01-26 17:02:33.959849989 +0000 UTC m=+249.913103970" observedRunningTime="2026-01-26 17:02:34.596784196 +0000 UTC m=+250.550038177" watchObservedRunningTime="2026-01-26 17:02:34.601280986 +0000 UTC m=+250.554534967" Jan 26 17:02:34 crc kubenswrapper[4856]: I0126 17:02:34.619736 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4kwt4" podStartSLOduration=2.718136475 podStartE2EDuration="1m15.619695987s" podCreationTimestamp="2026-01-26 17:01:19 +0000 UTC" firstStartedPulling="2026-01-26 17:01:21.081665711 +0000 UTC m=+177.034919692" lastFinishedPulling="2026-01-26 17:02:33.983225223 +0000 UTC m=+249.936479204" observedRunningTime="2026-01-26 17:02:34.617097442 +0000 UTC m=+250.570351443" watchObservedRunningTime="2026-01-26 17:02:34.619695987 +0000 UTC m=+250.572949968" Jan 26 17:02:35 crc kubenswrapper[4856]: I0126 17:02:35.585998 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-dnvfq" event={"ID":"26fc27be-aaf6-4ce1-9b90-0bca184e8f12","Type":"ContainerStarted","Data":"6ad5234eb21fa7846692d1d599422d832493eb58c2a0ff344f7f0e1e1fea0b14"} Jan 26 17:02:35 crc kubenswrapper[4856]: I0126 17:02:35.586321 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-dnvfq" Jan 26 17:02:35 crc kubenswrapper[4856]: I0126 17:02:35.617971 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-dnvfq" podStartSLOduration=2.617946525 podStartE2EDuration="2.617946525s" podCreationTimestamp="2026-01-26 17:02:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:02:35.610617953 +0000 UTC m=+251.563871944" watchObservedRunningTime="2026-01-26 17:02:35.617946525 +0000 UTC m=+251.571200506" Jan 26 17:02:37 crc kubenswrapper[4856]: I0126 17:02:37.598005 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g8bgt" event={"ID":"0d7eb7b8-63ae-493a-850b-0b9f3b42e927","Type":"ContainerStarted","Data":"3926e8ab5a46c920f3ea8cad2d006d4f4059fc6b8c475a7f6f3a22211a28d019"} Jan 26 17:02:37 crc kubenswrapper[4856]: I0126 17:02:37.816758 4856 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-txmdl" Jan 26 17:02:37 crc kubenswrapper[4856]: I0126 17:02:37.816845 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-txmdl" Jan 26 17:02:38 crc kubenswrapper[4856]: I0126 17:02:38.091202 4856 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-txmdl" Jan 26 17:02:38 crc kubenswrapper[4856]: I0126 17:02:38.193046 4856 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-62nhd" Jan 26 17:02:38 crc kubenswrapper[4856]: I0126 17:02:38.193315 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-62nhd" Jan 26 17:02:38 crc kubenswrapper[4856]: I0126 17:02:38.233212 4856 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-62nhd" Jan 26 17:02:38 crc kubenswrapper[4856]: I0126 17:02:38.605050 4856 generic.go:334] "Generic (PLEG): container finished" podID="0d7eb7b8-63ae-493a-850b-0b9f3b42e927" containerID="3926e8ab5a46c920f3ea8cad2d006d4f4059fc6b8c475a7f6f3a22211a28d019" exitCode=0 Jan 26 17:02:38 crc kubenswrapper[4856]: I0126 17:02:38.605160 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g8bgt" event={"ID":"0d7eb7b8-63ae-493a-850b-0b9f3b42e927","Type":"ContainerDied","Data":"3926e8ab5a46c920f3ea8cad2d006d4f4059fc6b8c475a7f6f3a22211a28d019"} Jan 26 17:02:38 crc kubenswrapper[4856]: I0126 17:02:38.696303 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-txmdl" Jan 26 17:02:39 crc kubenswrapper[4856]: I0126 17:02:39.652696 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-62nhd" Jan 26 17:02:39 crc kubenswrapper[4856]: I0126 17:02:39.894102 4856 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4kwt4" Jan 26 17:02:39 crc kubenswrapper[4856]: I0126 17:02:39.894397 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4kwt4" Jan 26 17:02:39 crc kubenswrapper[4856]: I0126 17:02:39.928165 4856 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4kwt4" Jan 26 17:02:40 crc kubenswrapper[4856]: I0126 17:02:40.655652 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4kwt4" Jan 26 17:02:40 crc kubenswrapper[4856]: I0126 17:02:40.967644 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-62nhd"] Jan 26 17:02:41 crc kubenswrapper[4856]: I0126 17:02:41.621481 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-62nhd" podUID="7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766" containerName="registry-server" containerID="cri-o://67c41d7af13d33af9423d069e86b531ff9d226b1435b62347517f490f3904943" gracePeriod=2 Jan 26 17:02:44 crc kubenswrapper[4856]: I0126 17:02:44.073555 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-txmdl"] Jan 26 17:02:44 crc kubenswrapper[4856]: I0126 17:02:44.074320 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-txmdl" podUID="40a27476-22b1-4083-990e-66e70ccdaf4c" containerName="registry-server" containerID="cri-o://d2eb8e794ba046c4faf481951340700b775e721a354e0ab5bea576c03e396810" gracePeriod=30 Jan 26 17:02:44 crc kubenswrapper[4856]: I0126 17:02:44.090955 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n8hp2"] Jan 26 17:02:44 crc kubenswrapper[4856]: I0126 17:02:44.097467 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qgjjd"] Jan 26 17:02:44 crc kubenswrapper[4856]: I0126 17:02:44.109910 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wvttb"] Jan 26 17:02:44 crc kubenswrapper[4856]: I0126 17:02:44.110368 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-wvttb" podUID="2d37efbf-d18f-486b-9b43-bc4d181af4ca" containerName="marketplace-operator" containerID="cri-o://743ebe09ef635c21a62370a80c15b76e3ff5e7e1801bb955f28ed30f848dcca9" gracePeriod=30 Jan 26 17:02:44 crc kubenswrapper[4856]: I0126 17:02:44.123218 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4kwt4"] Jan 26 17:02:44 crc kubenswrapper[4856]: I0126 17:02:44.123591 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4kwt4" podUID="d6944fc9-b8d7-4013-8702-b5765c410a0b" containerName="registry-server" containerID="cri-o://a6302eb5e39718f049dce88c0f2a26632538c4eb99b7dbea0dca8f7aae8306c7" gracePeriod=30 Jan 26 17:02:44 crc kubenswrapper[4856]: I0126 17:02:44.152865 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tdtfh"] Jan 26 17:02:44 crc kubenswrapper[4856]: I0126 17:02:44.175841 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-g8bgt"] Jan 26 17:02:44 crc kubenswrapper[4856]: I0126 17:02:44.176020 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-tdtfh" Jan 26 17:02:44 crc kubenswrapper[4856]: I0126 17:02:44.177486 4856 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-wvttb container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" start-of-body= Jan 26 17:02:44 crc kubenswrapper[4856]: I0126 17:02:44.177557 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-wvttb" podUID="2d37efbf-d18f-486b-9b43-bc4d181af4ca" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" Jan 26 17:02:44 crc kubenswrapper[4856]: I0126 17:02:44.181060 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mqxwf"] Jan 26 17:02:44 crc kubenswrapper[4856]: I0126 17:02:44.194666 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tdtfh"] Jan 26 17:02:44 crc kubenswrapper[4856]: I0126 17:02:44.196498 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qknj9"] Jan 26 17:02:44 crc kubenswrapper[4856]: I0126 17:02:44.275223 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/566ca894-037a-4b73-95d4-a6246c7c851a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-tdtfh\" (UID: \"566ca894-037a-4b73-95d4-a6246c7c851a\") " pod="openshift-marketplace/marketplace-operator-79b997595-tdtfh" Jan 26 17:02:44 crc kubenswrapper[4856]: I0126 17:02:44.275284 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmkjj\" (UniqueName: \"kubernetes.io/projected/566ca894-037a-4b73-95d4-a6246c7c851a-kube-api-access-wmkjj\") pod \"marketplace-operator-79b997595-tdtfh\" (UID: \"566ca894-037a-4b73-95d4-a6246c7c851a\") " pod="openshift-marketplace/marketplace-operator-79b997595-tdtfh" Jan 26 17:02:44 crc kubenswrapper[4856]: I0126 17:02:44.275353 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/566ca894-037a-4b73-95d4-a6246c7c851a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-tdtfh\" (UID: \"566ca894-037a-4b73-95d4-a6246c7c851a\") " pod="openshift-marketplace/marketplace-operator-79b997595-tdtfh" Jan 26 17:02:44 crc kubenswrapper[4856]: I0126 17:02:44.376589 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/566ca894-037a-4b73-95d4-a6246c7c851a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-tdtfh\" (UID: \"566ca894-037a-4b73-95d4-a6246c7c851a\") " pod="openshift-marketplace/marketplace-operator-79b997595-tdtfh" Jan 26 17:02:44 crc kubenswrapper[4856]: I0126 17:02:44.376971 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmkjj\" (UniqueName: \"kubernetes.io/projected/566ca894-037a-4b73-95d4-a6246c7c851a-kube-api-access-wmkjj\") pod \"marketplace-operator-79b997595-tdtfh\" (UID: \"566ca894-037a-4b73-95d4-a6246c7c851a\") " pod="openshift-marketplace/marketplace-operator-79b997595-tdtfh" Jan 26 17:02:44 crc kubenswrapper[4856]: I0126 17:02:44.377099 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/566ca894-037a-4b73-95d4-a6246c7c851a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-tdtfh\" (UID: \"566ca894-037a-4b73-95d4-a6246c7c851a\") " pod="openshift-marketplace/marketplace-operator-79b997595-tdtfh" Jan 26 17:02:44 crc kubenswrapper[4856]: I0126 17:02:44.377839 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/566ca894-037a-4b73-95d4-a6246c7c851a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-tdtfh\" (UID: \"566ca894-037a-4b73-95d4-a6246c7c851a\") " pod="openshift-marketplace/marketplace-operator-79b997595-tdtfh" Jan 26 17:02:44 crc kubenswrapper[4856]: I0126 17:02:44.382268 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/566ca894-037a-4b73-95d4-a6246c7c851a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-tdtfh\" (UID: \"566ca894-037a-4b73-95d4-a6246c7c851a\") " pod="openshift-marketplace/marketplace-operator-79b997595-tdtfh" Jan 26 17:02:44 crc kubenswrapper[4856]: I0126 17:02:44.397876 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmkjj\" (UniqueName: \"kubernetes.io/projected/566ca894-037a-4b73-95d4-a6246c7c851a-kube-api-access-wmkjj\") pod \"marketplace-operator-79b997595-tdtfh\" (UID: \"566ca894-037a-4b73-95d4-a6246c7c851a\") " pod="openshift-marketplace/marketplace-operator-79b997595-tdtfh" Jan 26 17:02:44 crc kubenswrapper[4856]: I0126 17:02:44.509393 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-tdtfh" Jan 26 17:02:44 crc kubenswrapper[4856]: I0126 17:02:44.626683 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-cb8nk"] Jan 26 17:02:46 crc kubenswrapper[4856]: I0126 17:02:46.648167 4856 generic.go:334] "Generic (PLEG): container finished" podID="2d37efbf-d18f-486b-9b43-bc4d181af4ca" containerID="743ebe09ef635c21a62370a80c15b76e3ff5e7e1801bb955f28ed30f848dcca9" exitCode=0 Jan 26 17:02:46 crc kubenswrapper[4856]: I0126 17:02:46.648260 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wvttb" event={"ID":"2d37efbf-d18f-486b-9b43-bc4d181af4ca","Type":"ContainerDied","Data":"743ebe09ef635c21a62370a80c15b76e3ff5e7e1801bb955f28ed30f848dcca9"} Jan 26 17:02:46 crc kubenswrapper[4856]: I0126 17:02:46.650781 4856 generic.go:334] "Generic (PLEG): container finished" podID="40a27476-22b1-4083-990e-66e70ccdaf4c" containerID="d2eb8e794ba046c4faf481951340700b775e721a354e0ab5bea576c03e396810" exitCode=0 Jan 26 17:02:46 crc kubenswrapper[4856]: I0126 17:02:46.650857 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-txmdl" event={"ID":"40a27476-22b1-4083-990e-66e70ccdaf4c","Type":"ContainerDied","Data":"d2eb8e794ba046c4faf481951340700b775e721a354e0ab5bea576c03e396810"} Jan 26 17:02:46 crc kubenswrapper[4856]: I0126 17:02:46.652376 4856 generic.go:334] "Generic (PLEG): container finished" podID="7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766" containerID="67c41d7af13d33af9423d069e86b531ff9d226b1435b62347517f490f3904943" exitCode=0 Jan 26 17:02:46 crc kubenswrapper[4856]: I0126 17:02:46.652446 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-62nhd" event={"ID":"7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766","Type":"ContainerDied","Data":"67c41d7af13d33af9423d069e86b531ff9d226b1435b62347517f490f3904943"} Jan 26 17:02:47 crc kubenswrapper[4856]: I0126 17:02:47.659300 4856 generic.go:334] "Generic (PLEG): container finished" podID="d6944fc9-b8d7-4013-8702-b5765c410a0b" containerID="a6302eb5e39718f049dce88c0f2a26632538c4eb99b7dbea0dca8f7aae8306c7" exitCode=0 Jan 26 17:02:47 crc kubenswrapper[4856]: I0126 17:02:47.659382 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4kwt4" event={"ID":"d6944fc9-b8d7-4013-8702-b5765c410a0b","Type":"ContainerDied","Data":"a6302eb5e39718f049dce88c0f2a26632538c4eb99b7dbea0dca8f7aae8306c7"} Jan 26 17:02:47 crc kubenswrapper[4856]: E0126 17:02:47.817232 4856 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d2eb8e794ba046c4faf481951340700b775e721a354e0ab5bea576c03e396810 is running failed: container process not found" containerID="d2eb8e794ba046c4faf481951340700b775e721a354e0ab5bea576c03e396810" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 17:02:47 crc kubenswrapper[4856]: E0126 17:02:47.817691 4856 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d2eb8e794ba046c4faf481951340700b775e721a354e0ab5bea576c03e396810 is running failed: container process not found" containerID="d2eb8e794ba046c4faf481951340700b775e721a354e0ab5bea576c03e396810" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 17:02:47 crc kubenswrapper[4856]: E0126 17:02:47.818130 4856 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d2eb8e794ba046c4faf481951340700b775e721a354e0ab5bea576c03e396810 is running failed: container process not found" containerID="d2eb8e794ba046c4faf481951340700b775e721a354e0ab5bea576c03e396810" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 17:02:47 crc kubenswrapper[4856]: E0126 17:02:47.818167 4856 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d2eb8e794ba046c4faf481951340700b775e721a354e0ab5bea576c03e396810 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-txmdl" podUID="40a27476-22b1-4083-990e-66e70ccdaf4c" containerName="registry-server" Jan 26 17:02:48 crc kubenswrapper[4856]: E0126 17:02:48.193442 4856 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 67c41d7af13d33af9423d069e86b531ff9d226b1435b62347517f490f3904943 is running failed: container process not found" containerID="67c41d7af13d33af9423d069e86b531ff9d226b1435b62347517f490f3904943" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 17:02:48 crc kubenswrapper[4856]: E0126 17:02:48.193816 4856 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 67c41d7af13d33af9423d069e86b531ff9d226b1435b62347517f490f3904943 is running failed: container process not found" containerID="67c41d7af13d33af9423d069e86b531ff9d226b1435b62347517f490f3904943" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 17:02:48 crc kubenswrapper[4856]: E0126 17:02:48.194319 4856 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 67c41d7af13d33af9423d069e86b531ff9d226b1435b62347517f490f3904943 is running failed: container process not found" containerID="67c41d7af13d33af9423d069e86b531ff9d226b1435b62347517f490f3904943" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 17:02:48 crc kubenswrapper[4856]: E0126 17:02:48.194357 4856 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 67c41d7af13d33af9423d069e86b531ff9d226b1435b62347517f490f3904943 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-62nhd" podUID="7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766" containerName="registry-server" Jan 26 17:02:49 crc kubenswrapper[4856]: E0126 17:02:49.895157 4856 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a6302eb5e39718f049dce88c0f2a26632538c4eb99b7dbea0dca8f7aae8306c7 is running failed: container process not found" containerID="a6302eb5e39718f049dce88c0f2a26632538c4eb99b7dbea0dca8f7aae8306c7" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 17:02:49 crc kubenswrapper[4856]: E0126 17:02:49.896078 4856 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a6302eb5e39718f049dce88c0f2a26632538c4eb99b7dbea0dca8f7aae8306c7 is running failed: container process not found" containerID="a6302eb5e39718f049dce88c0f2a26632538c4eb99b7dbea0dca8f7aae8306c7" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 17:02:49 crc kubenswrapper[4856]: E0126 17:02:49.896449 4856 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a6302eb5e39718f049dce88c0f2a26632538c4eb99b7dbea0dca8f7aae8306c7 is running failed: container process not found" containerID="a6302eb5e39718f049dce88c0f2a26632538c4eb99b7dbea0dca8f7aae8306c7" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 17:02:49 crc kubenswrapper[4856]: E0126 17:02:49.896487 4856 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a6302eb5e39718f049dce88c0f2a26632538c4eb99b7dbea0dca8f7aae8306c7 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-4kwt4" podUID="d6944fc9-b8d7-4013-8702-b5765c410a0b" containerName="registry-server" Jan 26 17:02:51 crc kubenswrapper[4856]: I0126 17:02:51.684567 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-62nhd" event={"ID":"7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766","Type":"ContainerDied","Data":"54783f51c7d33737624b9dffb5983a3ed107d31d30f3fd03ab73e5627dfd4bfd"} Jan 26 17:02:51 crc kubenswrapper[4856]: I0126 17:02:51.685007 4856 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="54783f51c7d33737624b9dffb5983a3ed107d31d30f3fd03ab73e5627dfd4bfd" Jan 26 17:02:51 crc kubenswrapper[4856]: I0126 17:02:51.685943 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-62nhd" Jan 26 17:02:51 crc kubenswrapper[4856]: I0126 17:02:51.790274 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s92lp\" (UniqueName: \"kubernetes.io/projected/7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766-kube-api-access-s92lp\") pod \"7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766\" (UID: \"7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766\") " Jan 26 17:02:51 crc kubenswrapper[4856]: I0126 17:02:51.790685 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766-utilities\") pod \"7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766\" (UID: \"7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766\") " Jan 26 17:02:51 crc kubenswrapper[4856]: I0126 17:02:51.790811 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766-catalog-content\") pod \"7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766\" (UID: \"7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766\") " Jan 26 17:02:51 crc kubenswrapper[4856]: I0126 17:02:51.792089 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766-utilities" (OuterVolumeSpecName: "utilities") pod "7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766" (UID: "7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:02:51 crc kubenswrapper[4856]: I0126 17:02:51.799858 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766-kube-api-access-s92lp" (OuterVolumeSpecName: "kube-api-access-s92lp") pod "7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766" (UID: "7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766"). InnerVolumeSpecName "kube-api-access-s92lp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:02:51 crc kubenswrapper[4856]: I0126 17:02:51.808284 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-txmdl" Jan 26 17:02:51 crc kubenswrapper[4856]: I0126 17:02:51.891942 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s92lp\" (UniqueName: \"kubernetes.io/projected/7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766-kube-api-access-s92lp\") on node \"crc\" DevicePath \"\"" Jan 26 17:02:51 crc kubenswrapper[4856]: I0126 17:02:51.891978 4856 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 17:02:51 crc kubenswrapper[4856]: I0126 17:02:51.897617 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wvttb" Jan 26 17:02:51 crc kubenswrapper[4856]: I0126 17:02:51.909963 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766" (UID: "7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:02:51 crc kubenswrapper[4856]: I0126 17:02:51.912647 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4kwt4" Jan 26 17:02:51 crc kubenswrapper[4856]: I0126 17:02:51.993011 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40a27476-22b1-4083-990e-66e70ccdaf4c-utilities\") pod \"40a27476-22b1-4083-990e-66e70ccdaf4c\" (UID: \"40a27476-22b1-4083-990e-66e70ccdaf4c\") " Jan 26 17:02:51 crc kubenswrapper[4856]: I0126 17:02:51.993131 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tf49w\" (UniqueName: \"kubernetes.io/projected/40a27476-22b1-4083-990e-66e70ccdaf4c-kube-api-access-tf49w\") pod \"40a27476-22b1-4083-990e-66e70ccdaf4c\" (UID: \"40a27476-22b1-4083-990e-66e70ccdaf4c\") " Jan 26 17:02:51 crc kubenswrapper[4856]: I0126 17:02:51.993171 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40a27476-22b1-4083-990e-66e70ccdaf4c-catalog-content\") pod \"40a27476-22b1-4083-990e-66e70ccdaf4c\" (UID: \"40a27476-22b1-4083-990e-66e70ccdaf4c\") " Jan 26 17:02:51 crc kubenswrapper[4856]: I0126 17:02:51.993423 4856 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 17:02:51 crc kubenswrapper[4856]: I0126 17:02:51.996221 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40a27476-22b1-4083-990e-66e70ccdaf4c-utilities" (OuterVolumeSpecName: "utilities") pod "40a27476-22b1-4083-990e-66e70ccdaf4c" (UID: "40a27476-22b1-4083-990e-66e70ccdaf4c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.006789 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40a27476-22b1-4083-990e-66e70ccdaf4c-kube-api-access-tf49w" (OuterVolumeSpecName: "kube-api-access-tf49w") pod "40a27476-22b1-4083-990e-66e70ccdaf4c" (UID: "40a27476-22b1-4083-990e-66e70ccdaf4c"). InnerVolumeSpecName "kube-api-access-tf49w". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.096395 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b8qsf\" (UniqueName: \"kubernetes.io/projected/2d37efbf-d18f-486b-9b43-bc4d181af4ca-kube-api-access-b8qsf\") pod \"2d37efbf-d18f-486b-9b43-bc4d181af4ca\" (UID: \"2d37efbf-d18f-486b-9b43-bc4d181af4ca\") " Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.098083 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6944fc9-b8d7-4013-8702-b5765c410a0b-catalog-content\") pod \"d6944fc9-b8d7-4013-8702-b5765c410a0b\" (UID: \"d6944fc9-b8d7-4013-8702-b5765c410a0b\") " Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.098277 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6944fc9-b8d7-4013-8702-b5765c410a0b-utilities\") pod \"d6944fc9-b8d7-4013-8702-b5765c410a0b\" (UID: \"d6944fc9-b8d7-4013-8702-b5765c410a0b\") " Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.098470 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2d37efbf-d18f-486b-9b43-bc4d181af4ca-marketplace-operator-metrics\") pod \"2d37efbf-d18f-486b-9b43-bc4d181af4ca\" (UID: \"2d37efbf-d18f-486b-9b43-bc4d181af4ca\") " Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.098741 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2qzgz\" (UniqueName: \"kubernetes.io/projected/d6944fc9-b8d7-4013-8702-b5765c410a0b-kube-api-access-2qzgz\") pod \"d6944fc9-b8d7-4013-8702-b5765c410a0b\" (UID: \"d6944fc9-b8d7-4013-8702-b5765c410a0b\") " Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.098874 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2d37efbf-d18f-486b-9b43-bc4d181af4ca-marketplace-trusted-ca\") pod \"2d37efbf-d18f-486b-9b43-bc4d181af4ca\" (UID: \"2d37efbf-d18f-486b-9b43-bc4d181af4ca\") " Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.304420 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40a27476-22b1-4083-990e-66e70ccdaf4c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "40a27476-22b1-4083-990e-66e70ccdaf4c" (UID: "40a27476-22b1-4083-990e-66e70ccdaf4c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.304696 4856 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40a27476-22b1-4083-990e-66e70ccdaf4c-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.304732 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tf49w\" (UniqueName: \"kubernetes.io/projected/40a27476-22b1-4083-990e-66e70ccdaf4c-kube-api-access-tf49w\") on node \"crc\" DevicePath \"\"" Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.306231 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6944fc9-b8d7-4013-8702-b5765c410a0b-utilities" (OuterVolumeSpecName: "utilities") pod "d6944fc9-b8d7-4013-8702-b5765c410a0b" (UID: "d6944fc9-b8d7-4013-8702-b5765c410a0b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.310367 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d37efbf-d18f-486b-9b43-bc4d181af4ca-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "2d37efbf-d18f-486b-9b43-bc4d181af4ca" (UID: "2d37efbf-d18f-486b-9b43-bc4d181af4ca"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.326710 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d37efbf-d18f-486b-9b43-bc4d181af4ca-kube-api-access-b8qsf" (OuterVolumeSpecName: "kube-api-access-b8qsf") pod "2d37efbf-d18f-486b-9b43-bc4d181af4ca" (UID: "2d37efbf-d18f-486b-9b43-bc4d181af4ca"). InnerVolumeSpecName "kube-api-access-b8qsf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.328773 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tdtfh"] Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.336453 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d37efbf-d18f-486b-9b43-bc4d181af4ca-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "2d37efbf-d18f-486b-9b43-bc4d181af4ca" (UID: "2d37efbf-d18f-486b-9b43-bc4d181af4ca"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.340280 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6944fc9-b8d7-4013-8702-b5765c410a0b-kube-api-access-2qzgz" (OuterVolumeSpecName: "kube-api-access-2qzgz") pod "d6944fc9-b8d7-4013-8702-b5765c410a0b" (UID: "d6944fc9-b8d7-4013-8702-b5765c410a0b"). InnerVolumeSpecName "kube-api-access-2qzgz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.354558 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6944fc9-b8d7-4013-8702-b5765c410a0b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d6944fc9-b8d7-4013-8702-b5765c410a0b" (UID: "d6944fc9-b8d7-4013-8702-b5765c410a0b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:02:52 crc kubenswrapper[4856]: W0126 17:02:52.372984 4856 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod566ca894_037a_4b73_95d4_a6246c7c851a.slice/crio-9cd1ed3014b88b989fd630041a498b328f9b0617c0cba5a762ce4e78425e361e WatchSource:0}: Error finding container 9cd1ed3014b88b989fd630041a498b328f9b0617c0cba5a762ce4e78425e361e: Status 404 returned error can't find the container with id 9cd1ed3014b88b989fd630041a498b328f9b0617c0cba5a762ce4e78425e361e Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.405769 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b8qsf\" (UniqueName: \"kubernetes.io/projected/2d37efbf-d18f-486b-9b43-bc4d181af4ca-kube-api-access-b8qsf\") on node \"crc\" DevicePath \"\"" Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.405881 4856 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6944fc9-b8d7-4013-8702-b5765c410a0b-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.405928 4856 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6944fc9-b8d7-4013-8702-b5765c410a0b-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.405960 4856 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2d37efbf-d18f-486b-9b43-bc4d181af4ca-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.406142 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2qzgz\" (UniqueName: \"kubernetes.io/projected/d6944fc9-b8d7-4013-8702-b5765c410a0b-kube-api-access-2qzgz\") on node \"crc\" DevicePath \"\"" Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.406223 4856 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2d37efbf-d18f-486b-9b43-bc4d181af4ca-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.406265 4856 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40a27476-22b1-4083-990e-66e70ccdaf4c-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.692477 4856 generic.go:334] "Generic (PLEG): container finished" podID="89cf05de-642b-4574-9f79-45e7a3d4afa3" containerID="ce1db845ee974faa07417a8ed669be7680c5b2f3c82683fabe5144e8c8d7d22c" exitCode=0 Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.692574 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qgjjd" event={"ID":"89cf05de-642b-4574-9f79-45e7a3d4afa3","Type":"ContainerDied","Data":"ce1db845ee974faa07417a8ed669be7680c5b2f3c82683fabe5144e8c8d7d22c"} Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.698453 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wvttb" event={"ID":"2d37efbf-d18f-486b-9b43-bc4d181af4ca","Type":"ContainerDied","Data":"fff8ee4c0db342e8c666d6319a47d7101521fb44435e8030d5a5dc565b0b6c44"} Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.698734 4856 scope.go:117] "RemoveContainer" containerID="743ebe09ef635c21a62370a80c15b76e3ff5e7e1801bb955f28ed30f848dcca9" Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.698929 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wvttb" Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.712379 4856 generic.go:334] "Generic (PLEG): container finished" podID="a6086d4b-faeb-4a12-8e6a-2a178dfe374c" containerID="28595b618ba5f672f05780cdefc1a37904c64c77c313fd5d1eade9c6ec61abec" exitCode=0 Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.712429 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n8hp2" event={"ID":"a6086d4b-faeb-4a12-8e6a-2a178dfe374c","Type":"ContainerDied","Data":"28595b618ba5f672f05780cdefc1a37904c64c77c313fd5d1eade9c6ec61abec"} Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.717253 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mqxwf" event={"ID":"9c71e219-35d7-4e1e-a371-3456dfd29e83","Type":"ContainerStarted","Data":"f26b43b06d9a9590dc23d990cf3b883949996d6147e4127f3313a0a700b7a8da"} Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.717447 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mqxwf" podUID="9c71e219-35d7-4e1e-a371-3456dfd29e83" containerName="extract-content" containerID="cri-o://f26b43b06d9a9590dc23d990cf3b883949996d6147e4127f3313a0a700b7a8da" gracePeriod=30 Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.746645 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-tdtfh" event={"ID":"566ca894-037a-4b73-95d4-a6246c7c851a","Type":"ContainerStarted","Data":"519d7b33053858d4024a6ccd49792f293a0c0e3d4961a9a11caf19ec554e5298"} Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.746694 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-tdtfh" event={"ID":"566ca894-037a-4b73-95d4-a6246c7c851a","Type":"ContainerStarted","Data":"9cd1ed3014b88b989fd630041a498b328f9b0617c0cba5a762ce4e78425e361e"} Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.746909 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wvttb"] Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.747050 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-tdtfh" Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.748075 4856 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-tdtfh container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.61:8080/healthz\": dial tcp 10.217.0.61:8080: connect: connection refused" start-of-body= Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.748114 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-tdtfh" podUID="566ca894-037a-4b73-95d4-a6246c7c851a" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.61:8080/healthz\": dial tcp 10.217.0.61:8080: connect: connection refused" Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.750428 4856 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wvttb"] Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.765025 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g8bgt" event={"ID":"0d7eb7b8-63ae-493a-850b-0b9f3b42e927","Type":"ContainerStarted","Data":"53b807aa482bf2d95f65ed65fcc51ebdaee0a2490bc9574ce63e9c46227c37e6"} Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.765198 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-g8bgt" podUID="0d7eb7b8-63ae-493a-850b-0b9f3b42e927" containerName="registry-server" containerID="cri-o://53b807aa482bf2d95f65ed65fcc51ebdaee0a2490bc9574ce63e9c46227c37e6" gracePeriod=30 Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.772110 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4kwt4" event={"ID":"d6944fc9-b8d7-4013-8702-b5765c410a0b","Type":"ContainerDied","Data":"c6a85642ee783cdf59dd26ba744cc42773e760d42354900c16ebdd5e8e9ec111"} Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.772116 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4kwt4" Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.772166 4856 scope.go:117] "RemoveContainer" containerID="a6302eb5e39718f049dce88c0f2a26632538c4eb99b7dbea0dca8f7aae8306c7" Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.776858 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-txmdl" Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.776864 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-txmdl" event={"ID":"40a27476-22b1-4083-990e-66e70ccdaf4c","Type":"ContainerDied","Data":"894929ba59d66c867404dc7094d1e4c1b977bab79b099140b34c889e7b66ae16"} Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.778789 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-62nhd" Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.779656 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qknj9" podUID="a3fa94fe-e4ad-4171-b853-89878dc61569" containerName="extract-content" containerID="cri-o://f974d02f853250cd240f8efa3661017c6f59e32afaba8f242e28a7789f1e0242" gracePeriod=30 Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.779978 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qknj9" event={"ID":"a3fa94fe-e4ad-4171-b853-89878dc61569","Type":"ContainerStarted","Data":"f974d02f853250cd240f8efa3661017c6f59e32afaba8f242e28a7789f1e0242"} Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.797756 4856 scope.go:117] "RemoveContainer" containerID="25c95d58185d3429ff473dbc3a21342905624b31dd17e96573eb140be4c2402c" Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.814121 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-tdtfh" podStartSLOduration=8.814093849 podStartE2EDuration="8.814093849s" podCreationTimestamp="2026-01-26 17:02:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:02:52.80995737 +0000 UTC m=+268.763211361" watchObservedRunningTime="2026-01-26 17:02:52.814093849 +0000 UTC m=+268.767347830" Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.847307 4856 scope.go:117] "RemoveContainer" containerID="d4ffeb43e14865bfef28f884de6e5301087c2d9158d7a77b0c10a8dfec7c7ce2" Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.849215 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-g8bgt" podStartSLOduration=3.358497503 podStartE2EDuration="1m33.849199617s" podCreationTimestamp="2026-01-26 17:01:19 +0000 UTC" firstStartedPulling="2026-01-26 17:01:21.09655267 +0000 UTC m=+177.049806651" lastFinishedPulling="2026-01-26 17:02:51.587254784 +0000 UTC m=+267.540508765" observedRunningTime="2026-01-26 17:02:52.844690517 +0000 UTC m=+268.797944518" watchObservedRunningTime="2026-01-26 17:02:52.849199617 +0000 UTC m=+268.802453598" Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.891211 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-62nhd"] Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.893582 4856 scope.go:117] "RemoveContainer" containerID="d2eb8e794ba046c4faf481951340700b775e721a354e0ab5bea576c03e396810" Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.898336 4856 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-62nhd"] Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.914429 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-txmdl"] Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.917617 4856 scope.go:117] "RemoveContainer" containerID="4be8cc185ffcb38acac0516b9ba74f7fa439552b3ff463ccf813f91341bce48c" Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.918418 4856 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-txmdl"] Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.929939 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4kwt4"] Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.934237 4856 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4kwt4"] Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.946507 4856 scope.go:117] "RemoveContainer" containerID="3ec09320bb48de5d8b6709469f0f84953408cf650f51d872373c21616d43f0de" Jan 26 17:02:52 crc kubenswrapper[4856]: I0126 17:02:52.960640 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qgjjd" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.116295 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gtxjz\" (UniqueName: \"kubernetes.io/projected/89cf05de-642b-4574-9f79-45e7a3d4afa3-kube-api-access-gtxjz\") pod \"89cf05de-642b-4574-9f79-45e7a3d4afa3\" (UID: \"89cf05de-642b-4574-9f79-45e7a3d4afa3\") " Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.116359 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89cf05de-642b-4574-9f79-45e7a3d4afa3-utilities\") pod \"89cf05de-642b-4574-9f79-45e7a3d4afa3\" (UID: \"89cf05de-642b-4574-9f79-45e7a3d4afa3\") " Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.116457 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89cf05de-642b-4574-9f79-45e7a3d4afa3-catalog-content\") pod \"89cf05de-642b-4574-9f79-45e7a3d4afa3\" (UID: \"89cf05de-642b-4574-9f79-45e7a3d4afa3\") " Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.117989 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89cf05de-642b-4574-9f79-45e7a3d4afa3-utilities" (OuterVolumeSpecName: "utilities") pod "89cf05de-642b-4574-9f79-45e7a3d4afa3" (UID: "89cf05de-642b-4574-9f79-45e7a3d4afa3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.122574 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89cf05de-642b-4574-9f79-45e7a3d4afa3-kube-api-access-gtxjz" (OuterVolumeSpecName: "kube-api-access-gtxjz") pod "89cf05de-642b-4574-9f79-45e7a3d4afa3" (UID: "89cf05de-642b-4574-9f79-45e7a3d4afa3"). InnerVolumeSpecName "kube-api-access-gtxjz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.148896 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n8hp2" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.189782 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89cf05de-642b-4574-9f79-45e7a3d4afa3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "89cf05de-642b-4574-9f79-45e7a3d4afa3" (UID: "89cf05de-642b-4574-9f79-45e7a3d4afa3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.217701 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gtxjz\" (UniqueName: \"kubernetes.io/projected/89cf05de-642b-4574-9f79-45e7a3d4afa3-kube-api-access-gtxjz\") on node \"crc\" DevicePath \"\"" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.217780 4856 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89cf05de-642b-4574-9f79-45e7a3d4afa3-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.217795 4856 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89cf05de-642b-4574-9f79-45e7a3d4afa3-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.283249 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mqxwf" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.318826 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7mvx\" (UniqueName: \"kubernetes.io/projected/a6086d4b-faeb-4a12-8e6a-2a178dfe374c-kube-api-access-x7mvx\") pod \"a6086d4b-faeb-4a12-8e6a-2a178dfe374c\" (UID: \"a6086d4b-faeb-4a12-8e6a-2a178dfe374c\") " Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.318938 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6086d4b-faeb-4a12-8e6a-2a178dfe374c-utilities\") pod \"a6086d4b-faeb-4a12-8e6a-2a178dfe374c\" (UID: \"a6086d4b-faeb-4a12-8e6a-2a178dfe374c\") " Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.319015 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6086d4b-faeb-4a12-8e6a-2a178dfe374c-catalog-content\") pod \"a6086d4b-faeb-4a12-8e6a-2a178dfe374c\" (UID: \"a6086d4b-faeb-4a12-8e6a-2a178dfe374c\") " Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.319933 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6086d4b-faeb-4a12-8e6a-2a178dfe374c-utilities" (OuterVolumeSpecName: "utilities") pod "a6086d4b-faeb-4a12-8e6a-2a178dfe374c" (UID: "a6086d4b-faeb-4a12-8e6a-2a178dfe374c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.322052 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6086d4b-faeb-4a12-8e6a-2a178dfe374c-kube-api-access-x7mvx" (OuterVolumeSpecName: "kube-api-access-x7mvx") pod "a6086d4b-faeb-4a12-8e6a-2a178dfe374c" (UID: "a6086d4b-faeb-4a12-8e6a-2a178dfe374c"). InnerVolumeSpecName "kube-api-access-x7mvx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.347025 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g8bgt_0d7eb7b8-63ae-493a-850b-0b9f3b42e927/registry-server/0.log" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.347828 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g8bgt" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.380367 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6086d4b-faeb-4a12-8e6a-2a178dfe374c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a6086d4b-faeb-4a12-8e6a-2a178dfe374c" (UID: "a6086d4b-faeb-4a12-8e6a-2a178dfe374c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.410392 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d37efbf-d18f-486b-9b43-bc4d181af4ca" path="/var/lib/kubelet/pods/2d37efbf-d18f-486b-9b43-bc4d181af4ca/volumes" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.411026 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40a27476-22b1-4083-990e-66e70ccdaf4c" path="/var/lib/kubelet/pods/40a27476-22b1-4083-990e-66e70ccdaf4c/volumes" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.411788 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766" path="/var/lib/kubelet/pods/7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766/volumes" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.413016 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6944fc9-b8d7-4013-8702-b5765c410a0b" path="/var/lib/kubelet/pods/d6944fc9-b8d7-4013-8702-b5765c410a0b/volumes" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.419602 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6swzw\" (UniqueName: \"kubernetes.io/projected/9c71e219-35d7-4e1e-a371-3456dfd29e83-kube-api-access-6swzw\") pod \"9c71e219-35d7-4e1e-a371-3456dfd29e83\" (UID: \"9c71e219-35d7-4e1e-a371-3456dfd29e83\") " Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.419690 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c71e219-35d7-4e1e-a371-3456dfd29e83-catalog-content\") pod \"9c71e219-35d7-4e1e-a371-3456dfd29e83\" (UID: \"9c71e219-35d7-4e1e-a371-3456dfd29e83\") " Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.419786 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c71e219-35d7-4e1e-a371-3456dfd29e83-utilities\") pod \"9c71e219-35d7-4e1e-a371-3456dfd29e83\" (UID: \"9c71e219-35d7-4e1e-a371-3456dfd29e83\") " Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.420071 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7mvx\" (UniqueName: \"kubernetes.io/projected/a6086d4b-faeb-4a12-8e6a-2a178dfe374c-kube-api-access-x7mvx\") on node \"crc\" DevicePath \"\"" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.420086 4856 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6086d4b-faeb-4a12-8e6a-2a178dfe374c-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.420096 4856 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6086d4b-faeb-4a12-8e6a-2a178dfe374c-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.420895 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c71e219-35d7-4e1e-a371-3456dfd29e83-utilities" (OuterVolumeSpecName: "utilities") pod "9c71e219-35d7-4e1e-a371-3456dfd29e83" (UID: "9c71e219-35d7-4e1e-a371-3456dfd29e83"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.423387 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c71e219-35d7-4e1e-a371-3456dfd29e83-kube-api-access-6swzw" (OuterVolumeSpecName: "kube-api-access-6swzw") pod "9c71e219-35d7-4e1e-a371-3456dfd29e83" (UID: "9c71e219-35d7-4e1e-a371-3456dfd29e83"). InnerVolumeSpecName "kube-api-access-6swzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.521055 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d7eb7b8-63ae-493a-850b-0b9f3b42e927-catalog-content\") pod \"0d7eb7b8-63ae-493a-850b-0b9f3b42e927\" (UID: \"0d7eb7b8-63ae-493a-850b-0b9f3b42e927\") " Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.521434 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d7eb7b8-63ae-493a-850b-0b9f3b42e927-utilities\") pod \"0d7eb7b8-63ae-493a-850b-0b9f3b42e927\" (UID: \"0d7eb7b8-63ae-493a-850b-0b9f3b42e927\") " Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.521481 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dcbvc\" (UniqueName: \"kubernetes.io/projected/0d7eb7b8-63ae-493a-850b-0b9f3b42e927-kube-api-access-dcbvc\") pod \"0d7eb7b8-63ae-493a-850b-0b9f3b42e927\" (UID: \"0d7eb7b8-63ae-493a-850b-0b9f3b42e927\") " Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.522117 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6swzw\" (UniqueName: \"kubernetes.io/projected/9c71e219-35d7-4e1e-a371-3456dfd29e83-kube-api-access-6swzw\") on node \"crc\" DevicePath \"\"" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.522163 4856 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c71e219-35d7-4e1e-a371-3456dfd29e83-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.522423 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d7eb7b8-63ae-493a-850b-0b9f3b42e927-utilities" (OuterVolumeSpecName: "utilities") pod "0d7eb7b8-63ae-493a-850b-0b9f3b42e927" (UID: "0d7eb7b8-63ae-493a-850b-0b9f3b42e927"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.524365 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d7eb7b8-63ae-493a-850b-0b9f3b42e927-kube-api-access-dcbvc" (OuterVolumeSpecName: "kube-api-access-dcbvc") pod "0d7eb7b8-63ae-493a-850b-0b9f3b42e927" (UID: "0d7eb7b8-63ae-493a-850b-0b9f3b42e927"). InnerVolumeSpecName "kube-api-access-dcbvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.546756 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d7eb7b8-63ae-493a-850b-0b9f3b42e927-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0d7eb7b8-63ae-493a-850b-0b9f3b42e927" (UID: "0d7eb7b8-63ae-493a-850b-0b9f3b42e927"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.600112 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c71e219-35d7-4e1e-a371-3456dfd29e83-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9c71e219-35d7-4e1e-a371-3456dfd29e83" (UID: "9c71e219-35d7-4e1e-a371-3456dfd29e83"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.623111 4856 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d7eb7b8-63ae-493a-850b-0b9f3b42e927-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.623159 4856 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c71e219-35d7-4e1e-a371-3456dfd29e83-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.623172 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dcbvc\" (UniqueName: \"kubernetes.io/projected/0d7eb7b8-63ae-493a-850b-0b9f3b42e927-kube-api-access-dcbvc\") on node \"crc\" DevicePath \"\"" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.623182 4856 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d7eb7b8-63ae-493a-850b-0b9f3b42e927-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.642860 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qknj9_a3fa94fe-e4ad-4171-b853-89878dc61569/extract-content/0.log" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.643817 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qknj9" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.785661 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-g8bgt_0d7eb7b8-63ae-493a-850b-0b9f3b42e927/registry-server/0.log" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.786282 4856 generic.go:334] "Generic (PLEG): container finished" podID="0d7eb7b8-63ae-493a-850b-0b9f3b42e927" containerID="53b807aa482bf2d95f65ed65fcc51ebdaee0a2490bc9574ce63e9c46227c37e6" exitCode=1 Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.786350 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g8bgt" event={"ID":"0d7eb7b8-63ae-493a-850b-0b9f3b42e927","Type":"ContainerDied","Data":"53b807aa482bf2d95f65ed65fcc51ebdaee0a2490bc9574ce63e9c46227c37e6"} Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.786382 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g8bgt" event={"ID":"0d7eb7b8-63ae-493a-850b-0b9f3b42e927","Type":"ContainerDied","Data":"eca9c93c5c35ce3c6c300c833124d2e0c4c40f4feaf4a45bd12b4eecdb2f116c"} Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.786403 4856 scope.go:117] "RemoveContainer" containerID="53b807aa482bf2d95f65ed65fcc51ebdaee0a2490bc9574ce63e9c46227c37e6" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.786557 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g8bgt" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.792067 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n8hp2" event={"ID":"a6086d4b-faeb-4a12-8e6a-2a178dfe374c","Type":"ContainerDied","Data":"61bc611402534dad5a09a8edd4e25038026dc1769890ea9d2407a69eb9c888af"} Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.792629 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n8hp2" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.794962 4856 generic.go:334] "Generic (PLEG): container finished" podID="9c71e219-35d7-4e1e-a371-3456dfd29e83" containerID="f26b43b06d9a9590dc23d990cf3b883949996d6147e4127f3313a0a700b7a8da" exitCode=0 Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.795036 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mqxwf" event={"ID":"9c71e219-35d7-4e1e-a371-3456dfd29e83","Type":"ContainerDied","Data":"f26b43b06d9a9590dc23d990cf3b883949996d6147e4127f3313a0a700b7a8da"} Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.795064 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mqxwf" event={"ID":"9c71e219-35d7-4e1e-a371-3456dfd29e83","Type":"ContainerDied","Data":"d09c4604a24ed1fd63afc114569ecaa6c0c08542e351c04817bb0f8a62c19b49"} Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.795151 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mqxwf" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.802686 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qknj9_a3fa94fe-e4ad-4171-b853-89878dc61569/extract-content/0.log" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.803205 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qknj9" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.803309 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qknj9" event={"ID":"a3fa94fe-e4ad-4171-b853-89878dc61569","Type":"ContainerDied","Data":"f974d02f853250cd240f8efa3661017c6f59e32afaba8f242e28a7789f1e0242"} Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.803147 4856 generic.go:334] "Generic (PLEG): container finished" podID="a3fa94fe-e4ad-4171-b853-89878dc61569" containerID="f974d02f853250cd240f8efa3661017c6f59e32afaba8f242e28a7789f1e0242" exitCode=2 Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.803768 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qknj9" event={"ID":"a3fa94fe-e4ad-4171-b853-89878dc61569","Type":"ContainerDied","Data":"083b0b52d78f857657f62965a6b3636eba0ff933ac74b23de919043206cf9046"} Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.805568 4856 scope.go:117] "RemoveContainer" containerID="3926e8ab5a46c920f3ea8cad2d006d4f4059fc6b8c475a7f6f3a22211a28d019" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.807405 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qgjjd" event={"ID":"89cf05de-642b-4574-9f79-45e7a3d4afa3","Type":"ContainerDied","Data":"e31b957fac8983059a89e5a7867c6294be7613d1c35b810e6c7face168eea509"} Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.807453 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qgjjd" Jan 26 17:02:53 crc kubenswrapper[4856]: I0126 17:02:53.811988 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-tdtfh" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.029829 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3fa94fe-e4ad-4171-b853-89878dc61569-catalog-content\") pod \"a3fa94fe-e4ad-4171-b853-89878dc61569\" (UID: \"a3fa94fe-e4ad-4171-b853-89878dc61569\") " Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.029844 4856 scope.go:117] "RemoveContainer" containerID="a9fe692a78995f7dad7ea556edacc772eb429ab92938195725add9a17bbe9e7c" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.029900 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvpzb\" (UniqueName: \"kubernetes.io/projected/a3fa94fe-e4ad-4171-b853-89878dc61569-kube-api-access-wvpzb\") pod \"a3fa94fe-e4ad-4171-b853-89878dc61569\" (UID: \"a3fa94fe-e4ad-4171-b853-89878dc61569\") " Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.029957 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3fa94fe-e4ad-4171-b853-89878dc61569-utilities\") pod \"a3fa94fe-e4ad-4171-b853-89878dc61569\" (UID: \"a3fa94fe-e4ad-4171-b853-89878dc61569\") " Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.030734 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3fa94fe-e4ad-4171-b853-89878dc61569-utilities" (OuterVolumeSpecName: "utilities") pod "a3fa94fe-e4ad-4171-b853-89878dc61569" (UID: "a3fa94fe-e4ad-4171-b853-89878dc61569"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.030824 4856 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a3fa94fe-e4ad-4171-b853-89878dc61569-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.050719 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3fa94fe-e4ad-4171-b853-89878dc61569-kube-api-access-wvpzb" (OuterVolumeSpecName: "kube-api-access-wvpzb") pod "a3fa94fe-e4ad-4171-b853-89878dc61569" (UID: "a3fa94fe-e4ad-4171-b853-89878dc61569"). InnerVolumeSpecName "kube-api-access-wvpzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.058927 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n8hp2"] Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.062981 4856 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-n8hp2"] Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.078458 4856 scope.go:117] "RemoveContainer" containerID="53b807aa482bf2d95f65ed65fcc51ebdaee0a2490bc9574ce63e9c46227c37e6" Jan 26 17:02:54 crc kubenswrapper[4856]: E0126 17:02:54.079104 4856 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53b807aa482bf2d95f65ed65fcc51ebdaee0a2490bc9574ce63e9c46227c37e6\": container with ID starting with 53b807aa482bf2d95f65ed65fcc51ebdaee0a2490bc9574ce63e9c46227c37e6 not found: ID does not exist" containerID="53b807aa482bf2d95f65ed65fcc51ebdaee0a2490bc9574ce63e9c46227c37e6" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.079146 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53b807aa482bf2d95f65ed65fcc51ebdaee0a2490bc9574ce63e9c46227c37e6"} err="failed to get container status \"53b807aa482bf2d95f65ed65fcc51ebdaee0a2490bc9574ce63e9c46227c37e6\": rpc error: code = NotFound desc = could not find container \"53b807aa482bf2d95f65ed65fcc51ebdaee0a2490bc9574ce63e9c46227c37e6\": container with ID starting with 53b807aa482bf2d95f65ed65fcc51ebdaee0a2490bc9574ce63e9c46227c37e6 not found: ID does not exist" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.079178 4856 scope.go:117] "RemoveContainer" containerID="3926e8ab5a46c920f3ea8cad2d006d4f4059fc6b8c475a7f6f3a22211a28d019" Jan 26 17:02:54 crc kubenswrapper[4856]: E0126 17:02:54.079601 4856 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3926e8ab5a46c920f3ea8cad2d006d4f4059fc6b8c475a7f6f3a22211a28d019\": container with ID starting with 3926e8ab5a46c920f3ea8cad2d006d4f4059fc6b8c475a7f6f3a22211a28d019 not found: ID does not exist" containerID="3926e8ab5a46c920f3ea8cad2d006d4f4059fc6b8c475a7f6f3a22211a28d019" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.079624 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3926e8ab5a46c920f3ea8cad2d006d4f4059fc6b8c475a7f6f3a22211a28d019"} err="failed to get container status \"3926e8ab5a46c920f3ea8cad2d006d4f4059fc6b8c475a7f6f3a22211a28d019\": rpc error: code = NotFound desc = could not find container \"3926e8ab5a46c920f3ea8cad2d006d4f4059fc6b8c475a7f6f3a22211a28d019\": container with ID starting with 3926e8ab5a46c920f3ea8cad2d006d4f4059fc6b8c475a7f6f3a22211a28d019 not found: ID does not exist" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.079636 4856 scope.go:117] "RemoveContainer" containerID="a9fe692a78995f7dad7ea556edacc772eb429ab92938195725add9a17bbe9e7c" Jan 26 17:02:54 crc kubenswrapper[4856]: E0126 17:02:54.079854 4856 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9fe692a78995f7dad7ea556edacc772eb429ab92938195725add9a17bbe9e7c\": container with ID starting with a9fe692a78995f7dad7ea556edacc772eb429ab92938195725add9a17bbe9e7c not found: ID does not exist" containerID="a9fe692a78995f7dad7ea556edacc772eb429ab92938195725add9a17bbe9e7c" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.079875 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9fe692a78995f7dad7ea556edacc772eb429ab92938195725add9a17bbe9e7c"} err="failed to get container status \"a9fe692a78995f7dad7ea556edacc772eb429ab92938195725add9a17bbe9e7c\": rpc error: code = NotFound desc = could not find container \"a9fe692a78995f7dad7ea556edacc772eb429ab92938195725add9a17bbe9e7c\": container with ID starting with a9fe692a78995f7dad7ea556edacc772eb429ab92938195725add9a17bbe9e7c not found: ID does not exist" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.079887 4856 scope.go:117] "RemoveContainer" containerID="28595b618ba5f672f05780cdefc1a37904c64c77c313fd5d1eade9c6ec61abec" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.091982 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mqxwf"] Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.107585 4856 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mqxwf"] Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.111328 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-dnvfq" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.113710 4856 scope.go:117] "RemoveContainer" containerID="5638f22e046bc8f28ee2834fa7820e942af58e17d4efe952168ca98e63b3fa12" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.116379 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-g8bgt"] Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.120860 4856 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-g8bgt"] Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.131893 4856 scope.go:117] "RemoveContainer" containerID="f26b43b06d9a9590dc23d990cf3b883949996d6147e4127f3313a0a700b7a8da" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.135427 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvpzb\" (UniqueName: \"kubernetes.io/projected/a3fa94fe-e4ad-4171-b853-89878dc61569-kube-api-access-wvpzb\") on node \"crc\" DevicePath \"\"" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.141303 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qgjjd"] Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.144288 4856 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-qgjjd"] Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.169135 4856 scope.go:117] "RemoveContainer" containerID="e6e9fc1c7474ee1cf14a50a96e79036f97d946d338d8a18c3434197cbd0438a8" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.190521 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3fa94fe-e4ad-4171-b853-89878dc61569-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a3fa94fe-e4ad-4171-b853-89878dc61569" (UID: "a3fa94fe-e4ad-4171-b853-89878dc61569"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.212508 4856 scope.go:117] "RemoveContainer" containerID="f26b43b06d9a9590dc23d990cf3b883949996d6147e4127f3313a0a700b7a8da" Jan 26 17:02:54 crc kubenswrapper[4856]: E0126 17:02:54.213905 4856 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f26b43b06d9a9590dc23d990cf3b883949996d6147e4127f3313a0a700b7a8da\": container with ID starting with f26b43b06d9a9590dc23d990cf3b883949996d6147e4127f3313a0a700b7a8da not found: ID does not exist" containerID="f26b43b06d9a9590dc23d990cf3b883949996d6147e4127f3313a0a700b7a8da" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.213970 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f26b43b06d9a9590dc23d990cf3b883949996d6147e4127f3313a0a700b7a8da"} err="failed to get container status \"f26b43b06d9a9590dc23d990cf3b883949996d6147e4127f3313a0a700b7a8da\": rpc error: code = NotFound desc = could not find container \"f26b43b06d9a9590dc23d990cf3b883949996d6147e4127f3313a0a700b7a8da\": container with ID starting with f26b43b06d9a9590dc23d990cf3b883949996d6147e4127f3313a0a700b7a8da not found: ID does not exist" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.214012 4856 scope.go:117] "RemoveContainer" containerID="e6e9fc1c7474ee1cf14a50a96e79036f97d946d338d8a18c3434197cbd0438a8" Jan 26 17:02:54 crc kubenswrapper[4856]: E0126 17:02:54.214385 4856 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6e9fc1c7474ee1cf14a50a96e79036f97d946d338d8a18c3434197cbd0438a8\": container with ID starting with e6e9fc1c7474ee1cf14a50a96e79036f97d946d338d8a18c3434197cbd0438a8 not found: ID does not exist" containerID="e6e9fc1c7474ee1cf14a50a96e79036f97d946d338d8a18c3434197cbd0438a8" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.214444 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6e9fc1c7474ee1cf14a50a96e79036f97d946d338d8a18c3434197cbd0438a8"} err="failed to get container status \"e6e9fc1c7474ee1cf14a50a96e79036f97d946d338d8a18c3434197cbd0438a8\": rpc error: code = NotFound desc = could not find container \"e6e9fc1c7474ee1cf14a50a96e79036f97d946d338d8a18c3434197cbd0438a8\": container with ID starting with e6e9fc1c7474ee1cf14a50a96e79036f97d946d338d8a18c3434197cbd0438a8 not found: ID does not exist" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.214480 4856 scope.go:117] "RemoveContainer" containerID="f974d02f853250cd240f8efa3661017c6f59e32afaba8f242e28a7789f1e0242" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.230674 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-wxbdh"] Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.236565 4856 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a3fa94fe-e4ad-4171-b853-89878dc61569-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.260824 4856 scope.go:117] "RemoveContainer" containerID="6c718aeedef34f07c2686370f8f78fe4060881e116396cc02bb806370cffdb47" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.281908 4856 scope.go:117] "RemoveContainer" containerID="f974d02f853250cd240f8efa3661017c6f59e32afaba8f242e28a7789f1e0242" Jan 26 17:02:54 crc kubenswrapper[4856]: E0126 17:02:54.282272 4856 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f974d02f853250cd240f8efa3661017c6f59e32afaba8f242e28a7789f1e0242\": container with ID starting with f974d02f853250cd240f8efa3661017c6f59e32afaba8f242e28a7789f1e0242 not found: ID does not exist" containerID="f974d02f853250cd240f8efa3661017c6f59e32afaba8f242e28a7789f1e0242" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.282314 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f974d02f853250cd240f8efa3661017c6f59e32afaba8f242e28a7789f1e0242"} err="failed to get container status \"f974d02f853250cd240f8efa3661017c6f59e32afaba8f242e28a7789f1e0242\": rpc error: code = NotFound desc = could not find container \"f974d02f853250cd240f8efa3661017c6f59e32afaba8f242e28a7789f1e0242\": container with ID starting with f974d02f853250cd240f8efa3661017c6f59e32afaba8f242e28a7789f1e0242 not found: ID does not exist" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.282342 4856 scope.go:117] "RemoveContainer" containerID="6c718aeedef34f07c2686370f8f78fe4060881e116396cc02bb806370cffdb47" Jan 26 17:02:54 crc kubenswrapper[4856]: E0126 17:02:54.282987 4856 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c718aeedef34f07c2686370f8f78fe4060881e116396cc02bb806370cffdb47\": container with ID starting with 6c718aeedef34f07c2686370f8f78fe4060881e116396cc02bb806370cffdb47 not found: ID does not exist" containerID="6c718aeedef34f07c2686370f8f78fe4060881e116396cc02bb806370cffdb47" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.283005 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c718aeedef34f07c2686370f8f78fe4060881e116396cc02bb806370cffdb47"} err="failed to get container status \"6c718aeedef34f07c2686370f8f78fe4060881e116396cc02bb806370cffdb47\": rpc error: code = NotFound desc = could not find container \"6c718aeedef34f07c2686370f8f78fe4060881e116396cc02bb806370cffdb47\": container with ID starting with 6c718aeedef34f07c2686370f8f78fe4060881e116396cc02bb806370cffdb47 not found: ID does not exist" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.283018 4856 scope.go:117] "RemoveContainer" containerID="ce1db845ee974faa07417a8ed669be7680c5b2f3c82683fabe5144e8c8d7d22c" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.301892 4856 scope.go:117] "RemoveContainer" containerID="de3e1fd7d5b6adab2150705e57df43577251e5278edb52956bb11f5539b1538a" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.373981 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bxhpt"] Jan 26 17:02:54 crc kubenswrapper[4856]: E0126 17:02:54.374261 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6944fc9-b8d7-4013-8702-b5765c410a0b" containerName="extract-utilities" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.374282 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6944fc9-b8d7-4013-8702-b5765c410a0b" containerName="extract-utilities" Jan 26 17:02:54 crc kubenswrapper[4856]: E0126 17:02:54.374295 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c71e219-35d7-4e1e-a371-3456dfd29e83" containerName="extract-content" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.374301 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c71e219-35d7-4e1e-a371-3456dfd29e83" containerName="extract-content" Jan 26 17:02:54 crc kubenswrapper[4856]: E0126 17:02:54.374312 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6944fc9-b8d7-4013-8702-b5765c410a0b" containerName="extract-content" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.374319 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6944fc9-b8d7-4013-8702-b5765c410a0b" containerName="extract-content" Jan 26 17:02:54 crc kubenswrapper[4856]: E0126 17:02:54.374326 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40a27476-22b1-4083-990e-66e70ccdaf4c" containerName="extract-content" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.374333 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="40a27476-22b1-4083-990e-66e70ccdaf4c" containerName="extract-content" Jan 26 17:02:54 crc kubenswrapper[4856]: E0126 17:02:54.374342 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40a27476-22b1-4083-990e-66e70ccdaf4c" containerName="registry-server" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.374347 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="40a27476-22b1-4083-990e-66e70ccdaf4c" containerName="registry-server" Jan 26 17:02:54 crc kubenswrapper[4856]: E0126 17:02:54.374355 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d37efbf-d18f-486b-9b43-bc4d181af4ca" containerName="marketplace-operator" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.374361 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d37efbf-d18f-486b-9b43-bc4d181af4ca" containerName="marketplace-operator" Jan 26 17:02:54 crc kubenswrapper[4856]: E0126 17:02:54.374367 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40a27476-22b1-4083-990e-66e70ccdaf4c" containerName="extract-utilities" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.374373 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="40a27476-22b1-4083-990e-66e70ccdaf4c" containerName="extract-utilities" Jan 26 17:02:54 crc kubenswrapper[4856]: E0126 17:02:54.374382 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3fa94fe-e4ad-4171-b853-89878dc61569" containerName="extract-utilities" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.374387 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3fa94fe-e4ad-4171-b853-89878dc61569" containerName="extract-utilities" Jan 26 17:02:54 crc kubenswrapper[4856]: E0126 17:02:54.374394 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6944fc9-b8d7-4013-8702-b5765c410a0b" containerName="registry-server" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.374399 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6944fc9-b8d7-4013-8702-b5765c410a0b" containerName="registry-server" Jan 26 17:02:54 crc kubenswrapper[4856]: E0126 17:02:54.374405 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766" containerName="extract-utilities" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.374413 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766" containerName="extract-utilities" Jan 26 17:02:54 crc kubenswrapper[4856]: E0126 17:02:54.374421 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6086d4b-faeb-4a12-8e6a-2a178dfe374c" containerName="extract-utilities" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.374427 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6086d4b-faeb-4a12-8e6a-2a178dfe374c" containerName="extract-utilities" Jan 26 17:02:54 crc kubenswrapper[4856]: E0126 17:02:54.374434 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c71e219-35d7-4e1e-a371-3456dfd29e83" containerName="extract-utilities" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.374440 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c71e219-35d7-4e1e-a371-3456dfd29e83" containerName="extract-utilities" Jan 26 17:02:54 crc kubenswrapper[4856]: E0126 17:02:54.374448 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766" containerName="registry-server" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.374454 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766" containerName="registry-server" Jan 26 17:02:54 crc kubenswrapper[4856]: E0126 17:02:54.374463 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89cf05de-642b-4574-9f79-45e7a3d4afa3" containerName="extract-utilities" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.374469 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="89cf05de-642b-4574-9f79-45e7a3d4afa3" containerName="extract-utilities" Jan 26 17:02:54 crc kubenswrapper[4856]: E0126 17:02:54.374477 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89cf05de-642b-4574-9f79-45e7a3d4afa3" containerName="extract-content" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.374482 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="89cf05de-642b-4574-9f79-45e7a3d4afa3" containerName="extract-content" Jan 26 17:02:54 crc kubenswrapper[4856]: E0126 17:02:54.374490 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6086d4b-faeb-4a12-8e6a-2a178dfe374c" containerName="extract-content" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.374495 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6086d4b-faeb-4a12-8e6a-2a178dfe374c" containerName="extract-content" Jan 26 17:02:54 crc kubenswrapper[4856]: E0126 17:02:54.374504 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d7eb7b8-63ae-493a-850b-0b9f3b42e927" containerName="extract-content" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.374509 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d7eb7b8-63ae-493a-850b-0b9f3b42e927" containerName="extract-content" Jan 26 17:02:54 crc kubenswrapper[4856]: E0126 17:02:54.374517 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3fa94fe-e4ad-4171-b853-89878dc61569" containerName="extract-content" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.374540 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3fa94fe-e4ad-4171-b853-89878dc61569" containerName="extract-content" Jan 26 17:02:54 crc kubenswrapper[4856]: E0126 17:02:54.374549 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d7eb7b8-63ae-493a-850b-0b9f3b42e927" containerName="registry-server" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.374554 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d7eb7b8-63ae-493a-850b-0b9f3b42e927" containerName="registry-server" Jan 26 17:02:54 crc kubenswrapper[4856]: E0126 17:02:54.374563 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d7eb7b8-63ae-493a-850b-0b9f3b42e927" containerName="extract-utilities" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.374570 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d7eb7b8-63ae-493a-850b-0b9f3b42e927" containerName="extract-utilities" Jan 26 17:02:54 crc kubenswrapper[4856]: E0126 17:02:54.374579 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766" containerName="extract-content" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.374585 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766" containerName="extract-content" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.374700 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6944fc9-b8d7-4013-8702-b5765c410a0b" containerName="registry-server" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.374711 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3fa94fe-e4ad-4171-b853-89878dc61569" containerName="extract-content" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.374719 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6086d4b-faeb-4a12-8e6a-2a178dfe374c" containerName="extract-content" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.374728 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fd0eddb-0f0f-4a37-b0b8-1d1b870a0766" containerName="registry-server" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.374734 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d7eb7b8-63ae-493a-850b-0b9f3b42e927" containerName="registry-server" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.374744 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d37efbf-d18f-486b-9b43-bc4d181af4ca" containerName="marketplace-operator" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.374750 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c71e219-35d7-4e1e-a371-3456dfd29e83" containerName="extract-content" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.374756 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="40a27476-22b1-4083-990e-66e70ccdaf4c" containerName="registry-server" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.374764 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="89cf05de-642b-4574-9f79-45e7a3d4afa3" containerName="extract-content" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.375680 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bxhpt" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.380319 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.387424 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bxhpt"] Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.445972 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qknj9"] Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.446490 4856 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qknj9"] Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.540354 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f34c6a8-6023-480c-a25e-46f8c4f3766b-catalog-content\") pod \"certified-operators-bxhpt\" (UID: \"5f34c6a8-6023-480c-a25e-46f8c4f3766b\") " pod="openshift-marketplace/certified-operators-bxhpt" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.540437 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crdtn\" (UniqueName: \"kubernetes.io/projected/5f34c6a8-6023-480c-a25e-46f8c4f3766b-kube-api-access-crdtn\") pod \"certified-operators-bxhpt\" (UID: \"5f34c6a8-6023-480c-a25e-46f8c4f3766b\") " pod="openshift-marketplace/certified-operators-bxhpt" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.540868 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f34c6a8-6023-480c-a25e-46f8c4f3766b-utilities\") pod \"certified-operators-bxhpt\" (UID: \"5f34c6a8-6023-480c-a25e-46f8c4f3766b\") " pod="openshift-marketplace/certified-operators-bxhpt" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.642311 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f34c6a8-6023-480c-a25e-46f8c4f3766b-utilities\") pod \"certified-operators-bxhpt\" (UID: \"5f34c6a8-6023-480c-a25e-46f8c4f3766b\") " pod="openshift-marketplace/certified-operators-bxhpt" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.642384 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f34c6a8-6023-480c-a25e-46f8c4f3766b-catalog-content\") pod \"certified-operators-bxhpt\" (UID: \"5f34c6a8-6023-480c-a25e-46f8c4f3766b\") " pod="openshift-marketplace/certified-operators-bxhpt" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.642432 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crdtn\" (UniqueName: \"kubernetes.io/projected/5f34c6a8-6023-480c-a25e-46f8c4f3766b-kube-api-access-crdtn\") pod \"certified-operators-bxhpt\" (UID: \"5f34c6a8-6023-480c-a25e-46f8c4f3766b\") " pod="openshift-marketplace/certified-operators-bxhpt" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.643261 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f34c6a8-6023-480c-a25e-46f8c4f3766b-catalog-content\") pod \"certified-operators-bxhpt\" (UID: \"5f34c6a8-6023-480c-a25e-46f8c4f3766b\") " pod="openshift-marketplace/certified-operators-bxhpt" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.643267 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f34c6a8-6023-480c-a25e-46f8c4f3766b-utilities\") pod \"certified-operators-bxhpt\" (UID: \"5f34c6a8-6023-480c-a25e-46f8c4f3766b\") " pod="openshift-marketplace/certified-operators-bxhpt" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.977000 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crdtn\" (UniqueName: \"kubernetes.io/projected/5f34c6a8-6023-480c-a25e-46f8c4f3766b-kube-api-access-crdtn\") pod \"certified-operators-bxhpt\" (UID: \"5f34c6a8-6023-480c-a25e-46f8c4f3766b\") " pod="openshift-marketplace/certified-operators-bxhpt" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.987926 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lfhpz"] Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.989105 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lfhpz" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.995038 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 26 17:02:54 crc kubenswrapper[4856]: I0126 17:02:54.995746 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bxhpt" Jan 26 17:02:55 crc kubenswrapper[4856]: I0126 17:02:55.002075 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lfhpz"] Jan 26 17:02:55 crc kubenswrapper[4856]: I0126 17:02:55.038561 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee-utilities\") pod \"redhat-operators-lfhpz\" (UID: \"8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee\") " pod="openshift-marketplace/redhat-operators-lfhpz" Jan 26 17:02:55 crc kubenswrapper[4856]: I0126 17:02:55.038630 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee-catalog-content\") pod \"redhat-operators-lfhpz\" (UID: \"8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee\") " pod="openshift-marketplace/redhat-operators-lfhpz" Jan 26 17:02:55 crc kubenswrapper[4856]: I0126 17:02:55.038705 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fm8n8\" (UniqueName: \"kubernetes.io/projected/8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee-kube-api-access-fm8n8\") pod \"redhat-operators-lfhpz\" (UID: \"8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee\") " pod="openshift-marketplace/redhat-operators-lfhpz" Jan 26 17:02:55 crc kubenswrapper[4856]: I0126 17:02:55.140309 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fm8n8\" (UniqueName: \"kubernetes.io/projected/8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee-kube-api-access-fm8n8\") pod \"redhat-operators-lfhpz\" (UID: \"8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee\") " pod="openshift-marketplace/redhat-operators-lfhpz" Jan 26 17:02:55 crc kubenswrapper[4856]: I0126 17:02:55.140906 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee-utilities\") pod \"redhat-operators-lfhpz\" (UID: \"8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee\") " pod="openshift-marketplace/redhat-operators-lfhpz" Jan 26 17:02:55 crc kubenswrapper[4856]: I0126 17:02:55.140963 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee-catalog-content\") pod \"redhat-operators-lfhpz\" (UID: \"8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee\") " pod="openshift-marketplace/redhat-operators-lfhpz" Jan 26 17:02:55 crc kubenswrapper[4856]: I0126 17:02:55.141616 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee-catalog-content\") pod \"redhat-operators-lfhpz\" (UID: \"8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee\") " pod="openshift-marketplace/redhat-operators-lfhpz" Jan 26 17:02:55 crc kubenswrapper[4856]: I0126 17:02:55.143319 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee-utilities\") pod \"redhat-operators-lfhpz\" (UID: \"8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee\") " pod="openshift-marketplace/redhat-operators-lfhpz" Jan 26 17:02:55 crc kubenswrapper[4856]: I0126 17:02:55.163281 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fm8n8\" (UniqueName: \"kubernetes.io/projected/8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee-kube-api-access-fm8n8\") pod \"redhat-operators-lfhpz\" (UID: \"8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee\") " pod="openshift-marketplace/redhat-operators-lfhpz" Jan 26 17:02:55 crc kubenswrapper[4856]: I0126 17:02:55.234817 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bxhpt"] Jan 26 17:02:55 crc kubenswrapper[4856]: I0126 17:02:55.312572 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lfhpz" Jan 26 17:02:55 crc kubenswrapper[4856]: I0126 17:02:55.423020 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d7eb7b8-63ae-493a-850b-0b9f3b42e927" path="/var/lib/kubelet/pods/0d7eb7b8-63ae-493a-850b-0b9f3b42e927/volumes" Jan 26 17:02:55 crc kubenswrapper[4856]: I0126 17:02:55.424918 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89cf05de-642b-4574-9f79-45e7a3d4afa3" path="/var/lib/kubelet/pods/89cf05de-642b-4574-9f79-45e7a3d4afa3/volumes" Jan 26 17:02:55 crc kubenswrapper[4856]: I0126 17:02:55.429538 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c71e219-35d7-4e1e-a371-3456dfd29e83" path="/var/lib/kubelet/pods/9c71e219-35d7-4e1e-a371-3456dfd29e83/volumes" Jan 26 17:02:55 crc kubenswrapper[4856]: I0126 17:02:55.430318 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3fa94fe-e4ad-4171-b853-89878dc61569" path="/var/lib/kubelet/pods/a3fa94fe-e4ad-4171-b853-89878dc61569/volumes" Jan 26 17:02:55 crc kubenswrapper[4856]: I0126 17:02:55.430940 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6086d4b-faeb-4a12-8e6a-2a178dfe374c" path="/var/lib/kubelet/pods/a6086d4b-faeb-4a12-8e6a-2a178dfe374c/volumes" Jan 26 17:02:55 crc kubenswrapper[4856]: I0126 17:02:55.715637 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lfhpz"] Jan 26 17:02:55 crc kubenswrapper[4856]: W0126 17:02:55.722437 4856 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8cd36133_7a25_4dae_83a4_bbd0fbf1f2ee.slice/crio-2ef5584084d8b1d348bc061565a055f6ec51467023159b7ea87a382eaa85c020 WatchSource:0}: Error finding container 2ef5584084d8b1d348bc061565a055f6ec51467023159b7ea87a382eaa85c020: Status 404 returned error can't find the container with id 2ef5584084d8b1d348bc061565a055f6ec51467023159b7ea87a382eaa85c020 Jan 26 17:02:55 crc kubenswrapper[4856]: I0126 17:02:55.969991 4856 generic.go:334] "Generic (PLEG): container finished" podID="8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee" containerID="88586be463e7344004a3be66277fc71e033d018b5c03dfa5c56597b48d237e72" exitCode=0 Jan 26 17:02:55 crc kubenswrapper[4856]: I0126 17:02:55.970082 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lfhpz" event={"ID":"8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee","Type":"ContainerDied","Data":"88586be463e7344004a3be66277fc71e033d018b5c03dfa5c56597b48d237e72"} Jan 26 17:02:55 crc kubenswrapper[4856]: I0126 17:02:55.970256 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lfhpz" event={"ID":"8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee","Type":"ContainerStarted","Data":"2ef5584084d8b1d348bc061565a055f6ec51467023159b7ea87a382eaa85c020"} Jan 26 17:02:55 crc kubenswrapper[4856]: I0126 17:02:55.974776 4856 generic.go:334] "Generic (PLEG): container finished" podID="5f34c6a8-6023-480c-a25e-46f8c4f3766b" containerID="405a5e16f820190605a6762d0b7653fa1a6bedd12b761afcd55093867a05ee57" exitCode=0 Jan 26 17:02:55 crc kubenswrapper[4856]: I0126 17:02:55.974821 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bxhpt" event={"ID":"5f34c6a8-6023-480c-a25e-46f8c4f3766b","Type":"ContainerDied","Data":"405a5e16f820190605a6762d0b7653fa1a6bedd12b761afcd55093867a05ee57"} Jan 26 17:02:55 crc kubenswrapper[4856]: I0126 17:02:55.974850 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bxhpt" event={"ID":"5f34c6a8-6023-480c-a25e-46f8c4f3766b","Type":"ContainerStarted","Data":"ae868d389f2d56b098915bbed54fc03534f4fd1519a0d344eda69f6356db31f0"} Jan 26 17:02:56 crc kubenswrapper[4856]: I0126 17:02:56.779142 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-97mff"] Jan 26 17:02:56 crc kubenswrapper[4856]: I0126 17:02:56.789071 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-97mff" Jan 26 17:02:56 crc kubenswrapper[4856]: I0126 17:02:56.793615 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-97mff"] Jan 26 17:02:56 crc kubenswrapper[4856]: I0126 17:02:56.794315 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 26 17:02:56 crc kubenswrapper[4856]: I0126 17:02:56.861861 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/886857c0-659b-4904-b75a-c55c3f712747-utilities\") pod \"redhat-marketplace-97mff\" (UID: \"886857c0-659b-4904-b75a-c55c3f712747\") " pod="openshift-marketplace/redhat-marketplace-97mff" Jan 26 17:02:56 crc kubenswrapper[4856]: I0126 17:02:56.861984 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9487\" (UniqueName: \"kubernetes.io/projected/886857c0-659b-4904-b75a-c55c3f712747-kube-api-access-q9487\") pod \"redhat-marketplace-97mff\" (UID: \"886857c0-659b-4904-b75a-c55c3f712747\") " pod="openshift-marketplace/redhat-marketplace-97mff" Jan 26 17:02:56 crc kubenswrapper[4856]: I0126 17:02:56.862034 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/886857c0-659b-4904-b75a-c55c3f712747-catalog-content\") pod \"redhat-marketplace-97mff\" (UID: \"886857c0-659b-4904-b75a-c55c3f712747\") " pod="openshift-marketplace/redhat-marketplace-97mff" Jan 26 17:02:56 crc kubenswrapper[4856]: I0126 17:02:56.963128 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/886857c0-659b-4904-b75a-c55c3f712747-catalog-content\") pod \"redhat-marketplace-97mff\" (UID: \"886857c0-659b-4904-b75a-c55c3f712747\") " pod="openshift-marketplace/redhat-marketplace-97mff" Jan 26 17:02:56 crc kubenswrapper[4856]: I0126 17:02:56.963242 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/886857c0-659b-4904-b75a-c55c3f712747-utilities\") pod \"redhat-marketplace-97mff\" (UID: \"886857c0-659b-4904-b75a-c55c3f712747\") " pod="openshift-marketplace/redhat-marketplace-97mff" Jan 26 17:02:56 crc kubenswrapper[4856]: I0126 17:02:56.963323 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9487\" (UniqueName: \"kubernetes.io/projected/886857c0-659b-4904-b75a-c55c3f712747-kube-api-access-q9487\") pod \"redhat-marketplace-97mff\" (UID: \"886857c0-659b-4904-b75a-c55c3f712747\") " pod="openshift-marketplace/redhat-marketplace-97mff" Jan 26 17:02:56 crc kubenswrapper[4856]: I0126 17:02:56.963817 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/886857c0-659b-4904-b75a-c55c3f712747-catalog-content\") pod \"redhat-marketplace-97mff\" (UID: \"886857c0-659b-4904-b75a-c55c3f712747\") " pod="openshift-marketplace/redhat-marketplace-97mff" Jan 26 17:02:56 crc kubenswrapper[4856]: I0126 17:02:56.964099 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/886857c0-659b-4904-b75a-c55c3f712747-utilities\") pod \"redhat-marketplace-97mff\" (UID: \"886857c0-659b-4904-b75a-c55c3f712747\") " pod="openshift-marketplace/redhat-marketplace-97mff" Jan 26 17:02:57 crc kubenswrapper[4856]: I0126 17:02:57.055436 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9487\" (UniqueName: \"kubernetes.io/projected/886857c0-659b-4904-b75a-c55c3f712747-kube-api-access-q9487\") pod \"redhat-marketplace-97mff\" (UID: \"886857c0-659b-4904-b75a-c55c3f712747\") " pod="openshift-marketplace/redhat-marketplace-97mff" Jan 26 17:02:57 crc kubenswrapper[4856]: I0126 17:02:57.115951 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-97mff" Jan 26 17:02:57 crc kubenswrapper[4856]: I0126 17:02:57.303018 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-97mff"] Jan 26 17:02:57 crc kubenswrapper[4856]: W0126 17:02:57.310753 4856 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod886857c0_659b_4904_b75a_c55c3f712747.slice/crio-2063535f537d4fe37e3e34708f04c20619c6cc50b85697e69a9333b26c91a793 WatchSource:0}: Error finding container 2063535f537d4fe37e3e34708f04c20619c6cc50b85697e69a9333b26c91a793: Status 404 returned error can't find the container with id 2063535f537d4fe37e3e34708f04c20619c6cc50b85697e69a9333b26c91a793 Jan 26 17:02:57 crc kubenswrapper[4856]: I0126 17:02:57.377451 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gdp2n"] Jan 26 17:02:57 crc kubenswrapper[4856]: I0126 17:02:57.379285 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gdp2n" Jan 26 17:02:57 crc kubenswrapper[4856]: I0126 17:02:57.382601 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 26 17:02:57 crc kubenswrapper[4856]: I0126 17:02:57.385564 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gdp2n"] Jan 26 17:02:57 crc kubenswrapper[4856]: I0126 17:02:57.571030 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4327b726-2edc-40ad-ac96-b19a7e020048-utilities\") pod \"community-operators-gdp2n\" (UID: \"4327b726-2edc-40ad-ac96-b19a7e020048\") " pod="openshift-marketplace/community-operators-gdp2n" Jan 26 17:02:57 crc kubenswrapper[4856]: I0126 17:02:57.571097 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4327b726-2edc-40ad-ac96-b19a7e020048-catalog-content\") pod \"community-operators-gdp2n\" (UID: \"4327b726-2edc-40ad-ac96-b19a7e020048\") " pod="openshift-marketplace/community-operators-gdp2n" Jan 26 17:02:57 crc kubenswrapper[4856]: I0126 17:02:57.571188 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tp8x9\" (UniqueName: \"kubernetes.io/projected/4327b726-2edc-40ad-ac96-b19a7e020048-kube-api-access-tp8x9\") pod \"community-operators-gdp2n\" (UID: \"4327b726-2edc-40ad-ac96-b19a7e020048\") " pod="openshift-marketplace/community-operators-gdp2n" Jan 26 17:02:57 crc kubenswrapper[4856]: I0126 17:02:57.672435 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4327b726-2edc-40ad-ac96-b19a7e020048-catalog-content\") pod \"community-operators-gdp2n\" (UID: \"4327b726-2edc-40ad-ac96-b19a7e020048\") " pod="openshift-marketplace/community-operators-gdp2n" Jan 26 17:02:57 crc kubenswrapper[4856]: I0126 17:02:57.672566 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tp8x9\" (UniqueName: \"kubernetes.io/projected/4327b726-2edc-40ad-ac96-b19a7e020048-kube-api-access-tp8x9\") pod \"community-operators-gdp2n\" (UID: \"4327b726-2edc-40ad-ac96-b19a7e020048\") " pod="openshift-marketplace/community-operators-gdp2n" Jan 26 17:02:57 crc kubenswrapper[4856]: I0126 17:02:57.672649 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4327b726-2edc-40ad-ac96-b19a7e020048-utilities\") pod \"community-operators-gdp2n\" (UID: \"4327b726-2edc-40ad-ac96-b19a7e020048\") " pod="openshift-marketplace/community-operators-gdp2n" Jan 26 17:02:57 crc kubenswrapper[4856]: I0126 17:02:57.673240 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4327b726-2edc-40ad-ac96-b19a7e020048-catalog-content\") pod \"community-operators-gdp2n\" (UID: \"4327b726-2edc-40ad-ac96-b19a7e020048\") " pod="openshift-marketplace/community-operators-gdp2n" Jan 26 17:02:57 crc kubenswrapper[4856]: I0126 17:02:57.673467 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4327b726-2edc-40ad-ac96-b19a7e020048-utilities\") pod \"community-operators-gdp2n\" (UID: \"4327b726-2edc-40ad-ac96-b19a7e020048\") " pod="openshift-marketplace/community-operators-gdp2n" Jan 26 17:02:57 crc kubenswrapper[4856]: I0126 17:02:57.696955 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tp8x9\" (UniqueName: \"kubernetes.io/projected/4327b726-2edc-40ad-ac96-b19a7e020048-kube-api-access-tp8x9\") pod \"community-operators-gdp2n\" (UID: \"4327b726-2edc-40ad-ac96-b19a7e020048\") " pod="openshift-marketplace/community-operators-gdp2n" Jan 26 17:02:57 crc kubenswrapper[4856]: I0126 17:02:57.810428 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gdp2n" Jan 26 17:02:58 crc kubenswrapper[4856]: I0126 17:02:58.108064 4856 generic.go:334] "Generic (PLEG): container finished" podID="5f34c6a8-6023-480c-a25e-46f8c4f3766b" containerID="ab4dc2168cf3030f71b121144c073ab78a14a965d8feaf1a0de933f786e1cb89" exitCode=0 Jan 26 17:02:58 crc kubenswrapper[4856]: I0126 17:02:58.108170 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bxhpt" event={"ID":"5f34c6a8-6023-480c-a25e-46f8c4f3766b","Type":"ContainerDied","Data":"ab4dc2168cf3030f71b121144c073ab78a14a965d8feaf1a0de933f786e1cb89"} Jan 26 17:02:58 crc kubenswrapper[4856]: I0126 17:02:58.111680 4856 generic.go:334] "Generic (PLEG): container finished" podID="886857c0-659b-4904-b75a-c55c3f712747" containerID="8cce484e79d411777eb43ce1a40864e7613f816cb566efdd41677d117f9c3633" exitCode=0 Jan 26 17:02:58 crc kubenswrapper[4856]: I0126 17:02:58.111731 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-97mff" event={"ID":"886857c0-659b-4904-b75a-c55c3f712747","Type":"ContainerDied","Data":"8cce484e79d411777eb43ce1a40864e7613f816cb566efdd41677d117f9c3633"} Jan 26 17:02:58 crc kubenswrapper[4856]: I0126 17:02:58.111756 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-97mff" event={"ID":"886857c0-659b-4904-b75a-c55c3f712747","Type":"ContainerStarted","Data":"2063535f537d4fe37e3e34708f04c20619c6cc50b85697e69a9333b26c91a793"} Jan 26 17:02:58 crc kubenswrapper[4856]: I0126 17:02:58.425918 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gdp2n"] Jan 26 17:02:59 crc kubenswrapper[4856]: I0126 17:02:59.120005 4856 generic.go:334] "Generic (PLEG): container finished" podID="4327b726-2edc-40ad-ac96-b19a7e020048" containerID="d5222989432010fc64c5d354d24c5e19cfaecc54d2025e00e9a6eb627c8732c1" exitCode=0 Jan 26 17:02:59 crc kubenswrapper[4856]: I0126 17:02:59.120078 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gdp2n" event={"ID":"4327b726-2edc-40ad-ac96-b19a7e020048","Type":"ContainerDied","Data":"d5222989432010fc64c5d354d24c5e19cfaecc54d2025e00e9a6eb627c8732c1"} Jan 26 17:02:59 crc kubenswrapper[4856]: I0126 17:02:59.120641 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gdp2n" event={"ID":"4327b726-2edc-40ad-ac96-b19a7e020048","Type":"ContainerStarted","Data":"d2be1441bb0d46a1b01561dfecdd886e3af79e4093d59f1a2017ef581ded6586"} Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.128313 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bxhpt" event={"ID":"5f34c6a8-6023-480c-a25e-46f8c4f3766b","Type":"ContainerStarted","Data":"094471a22cb2be3806f3e0c5d6e07e36d6a5a423b089fe73b28707ddde1dde10"} Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.130069 4856 generic.go:334] "Generic (PLEG): container finished" podID="886857c0-659b-4904-b75a-c55c3f712747" containerID="a75ef75367730507a8b7594226c5e9d4e14716073f574dda81c029b084dafd94" exitCode=0 Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.130334 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-97mff" event={"ID":"886857c0-659b-4904-b75a-c55c3f712747","Type":"ContainerDied","Data":"a75ef75367730507a8b7594226c5e9d4e14716073f574dda81c029b084dafd94"} Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.133244 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gdp2n" event={"ID":"4327b726-2edc-40ad-ac96-b19a7e020048","Type":"ContainerStarted","Data":"3c2c6909ed99198befbbb9ed59971fc2aac68a42311872081e29bb9546929cc6"} Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.151115 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bxhpt" podStartSLOduration=3.113900137 podStartE2EDuration="6.151097845s" podCreationTimestamp="2026-01-26 17:02:54 +0000 UTC" firstStartedPulling="2026-01-26 17:02:55.97615485 +0000 UTC m=+271.929408831" lastFinishedPulling="2026-01-26 17:02:59.013352558 +0000 UTC m=+274.966606539" observedRunningTime="2026-01-26 17:03:00.149702565 +0000 UTC m=+276.102956546" watchObservedRunningTime="2026-01-26 17:03:00.151097845 +0000 UTC m=+276.104351826" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.479645 4856 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.480803 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://89975b8f9428f81ab5d3fb48ced5dd9c837bea2feea3b89f5f7ff8d7d5d15b3e" gracePeriod=15 Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.481046 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://b03a01b651d9f66da4dd1f6e0d29ad97c0d6ae46b644c3d997d8dc99476706df" gracePeriod=15 Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.481116 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://f2daa3f13d37b7ae19dfca406b6b5cfdcd4f211287f7d410193f2fee36a24553" gracePeriod=15 Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.481172 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://40d432159a07ba20bf95f058bf8a597d67cbd8d852519bed035694f8ba3d8ec4" gracePeriod=15 Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.481213 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://59ba0f131eee14df389391e46bc772894e49e976c3663df5fb3bc98b3cb4a3d6" gracePeriod=15 Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.485557 4856 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 26 17:03:00 crc kubenswrapper[4856]: E0126 17:03:00.485818 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.485840 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 26 17:03:00 crc kubenswrapper[4856]: E0126 17:03:00.485856 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.485863 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 26 17:03:00 crc kubenswrapper[4856]: E0126 17:03:00.485873 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.485880 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 26 17:03:00 crc kubenswrapper[4856]: E0126 17:03:00.485901 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.485928 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 26 17:03:00 crc kubenswrapper[4856]: E0126 17:03:00.485936 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.485942 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Jan 26 17:03:00 crc kubenswrapper[4856]: E0126 17:03:00.485948 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.485955 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 26 17:03:00 crc kubenswrapper[4856]: E0126 17:03:00.485965 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.485970 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.486225 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.486245 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.486258 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.486266 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.486273 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.486283 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.486294 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 26 17:03:00 crc kubenswrapper[4856]: E0126 17:03:00.486405 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.486413 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.488475 4856 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.488944 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.489630 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.489739 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.489772 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.494107 4856 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.552991 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 26 17:03:00 crc kubenswrapper[4856]: E0126 17:03:00.574142 4856 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.241:6443: connect: connection refused" event="&Event{ObjectMeta:{redhat-marketplace-97mff.188e56a765f7220b openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:redhat-marketplace-97mff,UID:886857c0-659b-4904-b75a-c55c3f712747,APIVersion:v1,ResourceVersion:29932,FieldPath:spec.containers{registry-server},},Reason:Pulled,Message:Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\" in 441ms (442ms including waiting). Image size: 907837715 bytes.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-26 17:03:00.573405707 +0000 UTC m=+276.526659678,LastTimestamp:2026-01-26 17:03:00.573405707 +0000 UTC m=+276.526659678,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.594911 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.594964 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.594982 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.595018 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.595049 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.595068 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.595092 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.595106 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.595205 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.595237 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.595275 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.696295 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.696338 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.696377 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.696448 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.696483 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.696496 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.696570 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.696579 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.696633 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.696662 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 17:03:00 crc kubenswrapper[4856]: I0126 17:03:00.850688 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 17:03:01 crc kubenswrapper[4856]: I0126 17:03:01.141212 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-97mff" event={"ID":"886857c0-659b-4904-b75a-c55c3f712747","Type":"ContainerStarted","Data":"86283045d7d1049d9d8358f985c6aac8c275ef1f0b7a9715b13fd30bd1c328e7"} Jan 26 17:03:01 crc kubenswrapper[4856]: I0126 17:03:01.141823 4856 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:01 crc kubenswrapper[4856]: I0126 17:03:01.142046 4856 status_manager.go:851] "Failed to get status for pod" podUID="886857c0-659b-4904-b75a-c55c3f712747" pod="openshift-marketplace/redhat-marketplace-97mff" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-97mff\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:01 crc kubenswrapper[4856]: I0126 17:03:01.142071 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"57d0defac9d2de663e30454f59f9c50c448b069b057e3908291a344d9995f94b"} Jan 26 17:03:01 crc kubenswrapper[4856]: I0126 17:03:01.144143 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Jan 26 17:03:01 crc kubenswrapper[4856]: I0126 17:03:01.145433 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 26 17:03:01 crc kubenswrapper[4856]: I0126 17:03:01.146100 4856 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b03a01b651d9f66da4dd1f6e0d29ad97c0d6ae46b644c3d997d8dc99476706df" exitCode=0 Jan 26 17:03:01 crc kubenswrapper[4856]: I0126 17:03:01.146131 4856 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f2daa3f13d37b7ae19dfca406b6b5cfdcd4f211287f7d410193f2fee36a24553" exitCode=0 Jan 26 17:03:01 crc kubenswrapper[4856]: I0126 17:03:01.146145 4856 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="40d432159a07ba20bf95f058bf8a597d67cbd8d852519bed035694f8ba3d8ec4" exitCode=0 Jan 26 17:03:01 crc kubenswrapper[4856]: I0126 17:03:01.146154 4856 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="59ba0f131eee14df389391e46bc772894e49e976c3663df5fb3bc98b3cb4a3d6" exitCode=2 Jan 26 17:03:01 crc kubenswrapper[4856]: I0126 17:03:01.146134 4856 scope.go:117] "RemoveContainer" containerID="3f07438e20bdf71c752bb661084c835341999c561bb5442d75e177223881276f" Jan 26 17:03:01 crc kubenswrapper[4856]: I0126 17:03:01.150682 4856 generic.go:334] "Generic (PLEG): container finished" podID="4327b726-2edc-40ad-ac96-b19a7e020048" containerID="3c2c6909ed99198befbbb9ed59971fc2aac68a42311872081e29bb9546929cc6" exitCode=0 Jan 26 17:03:01 crc kubenswrapper[4856]: I0126 17:03:01.150768 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gdp2n" event={"ID":"4327b726-2edc-40ad-ac96-b19a7e020048","Type":"ContainerDied","Data":"3c2c6909ed99198befbbb9ed59971fc2aac68a42311872081e29bb9546929cc6"} Jan 26 17:03:01 crc kubenswrapper[4856]: I0126 17:03:01.151356 4856 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:01 crc kubenswrapper[4856]: I0126 17:03:01.151577 4856 status_manager.go:851] "Failed to get status for pod" podUID="4327b726-2edc-40ad-ac96-b19a7e020048" pod="openshift-marketplace/community-operators-gdp2n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-gdp2n\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:01 crc kubenswrapper[4856]: I0126 17:03:01.151753 4856 status_manager.go:851] "Failed to get status for pod" podUID="886857c0-659b-4904-b75a-c55c3f712747" pod="openshift-marketplace/redhat-marketplace-97mff" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-97mff\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:01 crc kubenswrapper[4856]: I0126 17:03:01.153502 4856 generic.go:334] "Generic (PLEG): container finished" podID="69379820-3062-4964-a8dd-8689f8cea38d" containerID="beeb8e8929ad597a53e5bcbe203dbd0aeea7fb6f4cfbcd350384cfbddded9459" exitCode=0 Jan 26 17:03:01 crc kubenswrapper[4856]: I0126 17:03:01.153997 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"69379820-3062-4964-a8dd-8689f8cea38d","Type":"ContainerDied","Data":"beeb8e8929ad597a53e5bcbe203dbd0aeea7fb6f4cfbcd350384cfbddded9459"} Jan 26 17:03:01 crc kubenswrapper[4856]: I0126 17:03:01.154390 4856 status_manager.go:851] "Failed to get status for pod" podUID="69379820-3062-4964-a8dd-8689f8cea38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:01 crc kubenswrapper[4856]: I0126 17:03:01.154842 4856 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:01 crc kubenswrapper[4856]: I0126 17:03:01.155025 4856 status_manager.go:851] "Failed to get status for pod" podUID="4327b726-2edc-40ad-ac96-b19a7e020048" pod="openshift-marketplace/community-operators-gdp2n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-gdp2n\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:01 crc kubenswrapper[4856]: I0126 17:03:01.155167 4856 status_manager.go:851] "Failed to get status for pod" podUID="886857c0-659b-4904-b75a-c55c3f712747" pod="openshift-marketplace/redhat-marketplace-97mff" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-97mff\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:02 crc kubenswrapper[4856]: I0126 17:03:02.160851 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"b6f864463a7443541f4490597f84a8f03e6d8c1a587e47002bb795632c0df2d6"} Jan 26 17:03:02 crc kubenswrapper[4856]: I0126 17:03:02.162397 4856 status_manager.go:851] "Failed to get status for pod" podUID="69379820-3062-4964-a8dd-8689f8cea38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:02 crc kubenswrapper[4856]: I0126 17:03:02.162795 4856 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:02 crc kubenswrapper[4856]: I0126 17:03:02.163191 4856 status_manager.go:851] "Failed to get status for pod" podUID="4327b726-2edc-40ad-ac96-b19a7e020048" pod="openshift-marketplace/community-operators-gdp2n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-gdp2n\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:02 crc kubenswrapper[4856]: I0126 17:03:02.163414 4856 status_manager.go:851] "Failed to get status for pod" podUID="886857c0-659b-4904-b75a-c55c3f712747" pod="openshift-marketplace/redhat-marketplace-97mff" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-97mff\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:02 crc kubenswrapper[4856]: I0126 17:03:02.164855 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 26 17:03:02 crc kubenswrapper[4856]: I0126 17:03:02.167901 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gdp2n" event={"ID":"4327b726-2edc-40ad-ac96-b19a7e020048","Type":"ContainerStarted","Data":"c8e6dd9ff8ca391ce4dc51ffe3ea5566118b1b3b87a7744e96837eab0a37a59a"} Jan 26 17:03:02 crc kubenswrapper[4856]: I0126 17:03:02.168752 4856 status_manager.go:851] "Failed to get status for pod" podUID="886857c0-659b-4904-b75a-c55c3f712747" pod="openshift-marketplace/redhat-marketplace-97mff" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-97mff\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:02 crc kubenswrapper[4856]: I0126 17:03:02.169232 4856 status_manager.go:851] "Failed to get status for pod" podUID="69379820-3062-4964-a8dd-8689f8cea38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:02 crc kubenswrapper[4856]: I0126 17:03:02.169615 4856 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:02 crc kubenswrapper[4856]: I0126 17:03:02.169826 4856 status_manager.go:851] "Failed to get status for pod" podUID="4327b726-2edc-40ad-ac96-b19a7e020048" pod="openshift-marketplace/community-operators-gdp2n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-gdp2n\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:02 crc kubenswrapper[4856]: I0126 17:03:02.480843 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 26 17:03:02 crc kubenswrapper[4856]: I0126 17:03:02.482133 4856 status_manager.go:851] "Failed to get status for pod" podUID="886857c0-659b-4904-b75a-c55c3f712747" pod="openshift-marketplace/redhat-marketplace-97mff" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-97mff\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:02 crc kubenswrapper[4856]: I0126 17:03:02.482668 4856 status_manager.go:851] "Failed to get status for pod" podUID="69379820-3062-4964-a8dd-8689f8cea38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:02 crc kubenswrapper[4856]: I0126 17:03:02.483030 4856 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:02 crc kubenswrapper[4856]: I0126 17:03:02.483471 4856 status_manager.go:851] "Failed to get status for pod" podUID="4327b726-2edc-40ad-ac96-b19a7e020048" pod="openshift-marketplace/community-operators-gdp2n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-gdp2n\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:02 crc kubenswrapper[4856]: I0126 17:03:02.623008 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/69379820-3062-4964-a8dd-8689f8cea38d-kubelet-dir\") pod \"69379820-3062-4964-a8dd-8689f8cea38d\" (UID: \"69379820-3062-4964-a8dd-8689f8cea38d\") " Jan 26 17:03:02 crc kubenswrapper[4856]: I0126 17:03:02.623466 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/69379820-3062-4964-a8dd-8689f8cea38d-kube-api-access\") pod \"69379820-3062-4964-a8dd-8689f8cea38d\" (UID: \"69379820-3062-4964-a8dd-8689f8cea38d\") " Jan 26 17:03:02 crc kubenswrapper[4856]: I0126 17:03:02.623498 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/69379820-3062-4964-a8dd-8689f8cea38d-var-lock\") pod \"69379820-3062-4964-a8dd-8689f8cea38d\" (UID: \"69379820-3062-4964-a8dd-8689f8cea38d\") " Jan 26 17:03:03 crc kubenswrapper[4856]: I0126 17:03:02.623920 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/69379820-3062-4964-a8dd-8689f8cea38d-var-lock" (OuterVolumeSpecName: "var-lock") pod "69379820-3062-4964-a8dd-8689f8cea38d" (UID: "69379820-3062-4964-a8dd-8689f8cea38d"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:03:03 crc kubenswrapper[4856]: I0126 17:03:02.623937 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/69379820-3062-4964-a8dd-8689f8cea38d-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "69379820-3062-4964-a8dd-8689f8cea38d" (UID: "69379820-3062-4964-a8dd-8689f8cea38d"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:03:03 crc kubenswrapper[4856]: I0126 17:03:02.629837 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69379820-3062-4964-a8dd-8689f8cea38d-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "69379820-3062-4964-a8dd-8689f8cea38d" (UID: "69379820-3062-4964-a8dd-8689f8cea38d"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:03:03 crc kubenswrapper[4856]: I0126 17:03:02.725137 4856 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/69379820-3062-4964-a8dd-8689f8cea38d-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 26 17:03:03 crc kubenswrapper[4856]: I0126 17:03:02.725168 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/69379820-3062-4964-a8dd-8689f8cea38d-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 26 17:03:03 crc kubenswrapper[4856]: I0126 17:03:02.725179 4856 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/69379820-3062-4964-a8dd-8689f8cea38d-var-lock\") on node \"crc\" DevicePath \"\"" Jan 26 17:03:03 crc kubenswrapper[4856]: I0126 17:03:03.178438 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"69379820-3062-4964-a8dd-8689f8cea38d","Type":"ContainerDied","Data":"af60ab5d4a2b57ad1bbcc4a879fdc9dce5f1b3ef1e2f5eb96e13241cdf6f2277"} Jan 26 17:03:03 crc kubenswrapper[4856]: I0126 17:03:03.178737 4856 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af60ab5d4a2b57ad1bbcc4a879fdc9dce5f1b3ef1e2f5eb96e13241cdf6f2277" Jan 26 17:03:03 crc kubenswrapper[4856]: I0126 17:03:03.178804 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 26 17:03:03 crc kubenswrapper[4856]: I0126 17:03:03.188062 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 26 17:03:03 crc kubenswrapper[4856]: I0126 17:03:03.189085 4856 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="89975b8f9428f81ab5d3fb48ced5dd9c837bea2feea3b89f5f7ff8d7d5d15b3e" exitCode=0 Jan 26 17:03:04 crc kubenswrapper[4856]: I0126 17:03:04.209999 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 26 17:03:04 crc kubenswrapper[4856]: I0126 17:03:04.214179 4856 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4365b662a1ef1f25c43b0d9068b29f4b8c92282da9679a062ca15b8955aa46e5" Jan 26 17:03:04 crc kubenswrapper[4856]: I0126 17:03:04.574418 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 26 17:03:04 crc kubenswrapper[4856]: I0126 17:03:04.575969 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 17:03:04 crc kubenswrapper[4856]: I0126 17:03:04.576980 4856 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:04 crc kubenswrapper[4856]: I0126 17:03:04.577724 4856 status_manager.go:851] "Failed to get status for pod" podUID="69379820-3062-4964-a8dd-8689f8cea38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:04 crc kubenswrapper[4856]: I0126 17:03:04.578063 4856 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:04 crc kubenswrapper[4856]: I0126 17:03:04.578248 4856 status_manager.go:851] "Failed to get status for pod" podUID="4327b726-2edc-40ad-ac96-b19a7e020048" pod="openshift-marketplace/community-operators-gdp2n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-gdp2n\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:04 crc kubenswrapper[4856]: I0126 17:03:04.578422 4856 status_manager.go:851] "Failed to get status for pod" podUID="886857c0-659b-4904-b75a-c55c3f712747" pod="openshift-marketplace/redhat-marketplace-97mff" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-97mff\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:04 crc kubenswrapper[4856]: E0126 17:03:04.591249 4856 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:04 crc kubenswrapper[4856]: E0126 17:03:04.591850 4856 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:04 crc kubenswrapper[4856]: E0126 17:03:04.592454 4856 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:04 crc kubenswrapper[4856]: E0126 17:03:04.592789 4856 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:04 crc kubenswrapper[4856]: E0126 17:03:04.593163 4856 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:04 crc kubenswrapper[4856]: I0126 17:03:04.593213 4856 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Jan 26 17:03:04 crc kubenswrapper[4856]: E0126 17:03:04.593704 4856 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.241:6443: connect: connection refused" interval="200ms" Jan 26 17:03:04 crc kubenswrapper[4856]: I0126 17:03:04.754053 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:03:04 crc kubenswrapper[4856]: I0126 17:03:04.754397 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 26 17:03:04 crc kubenswrapper[4856]: I0126 17:03:04.754537 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 26 17:03:04 crc kubenswrapper[4856]: I0126 17:03:04.754630 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:03:04 crc kubenswrapper[4856]: I0126 17:03:04.754712 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 26 17:03:04 crc kubenswrapper[4856]: I0126 17:03:04.754794 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:03:04 crc kubenswrapper[4856]: I0126 17:03:04.755199 4856 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Jan 26 17:03:04 crc kubenswrapper[4856]: I0126 17:03:04.755224 4856 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Jan 26 17:03:04 crc kubenswrapper[4856]: I0126 17:03:04.755236 4856 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 26 17:03:04 crc kubenswrapper[4856]: E0126 17:03:04.795756 4856 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.241:6443: connect: connection refused" interval="400ms" Jan 26 17:03:04 crc kubenswrapper[4856]: I0126 17:03:04.997090 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bxhpt" Jan 26 17:03:04 crc kubenswrapper[4856]: I0126 17:03:04.997154 4856 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bxhpt" Jan 26 17:03:05 crc kubenswrapper[4856]: I0126 17:03:05.038624 4856 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bxhpt" Jan 26 17:03:05 crc kubenswrapper[4856]: I0126 17:03:05.039254 4856 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:05 crc kubenswrapper[4856]: I0126 17:03:05.039741 4856 status_manager.go:851] "Failed to get status for pod" podUID="69379820-3062-4964-a8dd-8689f8cea38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:05 crc kubenswrapper[4856]: I0126 17:03:05.040096 4856 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:05 crc kubenswrapper[4856]: I0126 17:03:05.040485 4856 status_manager.go:851] "Failed to get status for pod" podUID="4327b726-2edc-40ad-ac96-b19a7e020048" pod="openshift-marketplace/community-operators-gdp2n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-gdp2n\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:05 crc kubenswrapper[4856]: I0126 17:03:05.040746 4856 status_manager.go:851] "Failed to get status for pod" podUID="5f34c6a8-6023-480c-a25e-46f8c4f3766b" pod="openshift-marketplace/certified-operators-bxhpt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-bxhpt\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:05 crc kubenswrapper[4856]: I0126 17:03:05.041029 4856 status_manager.go:851] "Failed to get status for pod" podUID="886857c0-659b-4904-b75a-c55c3f712747" pod="openshift-marketplace/redhat-marketplace-97mff" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-97mff\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:05 crc kubenswrapper[4856]: E0126 17:03:05.197413 4856 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.241:6443: connect: connection refused" interval="800ms" Jan 26 17:03:05 crc kubenswrapper[4856]: I0126 17:03:05.219126 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 17:03:05 crc kubenswrapper[4856]: I0126 17:03:05.236018 4856 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:05 crc kubenswrapper[4856]: I0126 17:03:05.236421 4856 status_manager.go:851] "Failed to get status for pod" podUID="4327b726-2edc-40ad-ac96-b19a7e020048" pod="openshift-marketplace/community-operators-gdp2n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-gdp2n\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:05 crc kubenswrapper[4856]: I0126 17:03:05.236959 4856 status_manager.go:851] "Failed to get status for pod" podUID="5f34c6a8-6023-480c-a25e-46f8c4f3766b" pod="openshift-marketplace/certified-operators-bxhpt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-bxhpt\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:05 crc kubenswrapper[4856]: I0126 17:03:05.237237 4856 status_manager.go:851] "Failed to get status for pod" podUID="886857c0-659b-4904-b75a-c55c3f712747" pod="openshift-marketplace/redhat-marketplace-97mff" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-97mff\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:05 crc kubenswrapper[4856]: I0126 17:03:05.237553 4856 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:05 crc kubenswrapper[4856]: I0126 17:03:05.237888 4856 status_manager.go:851] "Failed to get status for pod" podUID="69379820-3062-4964-a8dd-8689f8cea38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:05 crc kubenswrapper[4856]: I0126 17:03:05.257763 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bxhpt" Jan 26 17:03:05 crc kubenswrapper[4856]: I0126 17:03:05.258302 4856 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:05 crc kubenswrapper[4856]: I0126 17:03:05.258789 4856 status_manager.go:851] "Failed to get status for pod" podUID="4327b726-2edc-40ad-ac96-b19a7e020048" pod="openshift-marketplace/community-operators-gdp2n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-gdp2n\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:05 crc kubenswrapper[4856]: I0126 17:03:05.259330 4856 status_manager.go:851] "Failed to get status for pod" podUID="5f34c6a8-6023-480c-a25e-46f8c4f3766b" pod="openshift-marketplace/certified-operators-bxhpt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-bxhpt\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:05 crc kubenswrapper[4856]: I0126 17:03:05.259789 4856 status_manager.go:851] "Failed to get status for pod" podUID="886857c0-659b-4904-b75a-c55c3f712747" pod="openshift-marketplace/redhat-marketplace-97mff" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-97mff\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:05 crc kubenswrapper[4856]: I0126 17:03:05.260097 4856 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:05 crc kubenswrapper[4856]: I0126 17:03:05.260624 4856 status_manager.go:851] "Failed to get status for pod" podUID="69379820-3062-4964-a8dd-8689f8cea38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:05 crc kubenswrapper[4856]: I0126 17:03:05.397699 4856 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:05 crc kubenswrapper[4856]: I0126 17:03:05.398092 4856 status_manager.go:851] "Failed to get status for pod" podUID="69379820-3062-4964-a8dd-8689f8cea38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:05 crc kubenswrapper[4856]: I0126 17:03:05.398396 4856 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:05 crc kubenswrapper[4856]: I0126 17:03:05.398707 4856 status_manager.go:851] "Failed to get status for pod" podUID="4327b726-2edc-40ad-ac96-b19a7e020048" pod="openshift-marketplace/community-operators-gdp2n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-gdp2n\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:05 crc kubenswrapper[4856]: I0126 17:03:05.398937 4856 status_manager.go:851] "Failed to get status for pod" podUID="5f34c6a8-6023-480c-a25e-46f8c4f3766b" pod="openshift-marketplace/certified-operators-bxhpt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-bxhpt\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:05 crc kubenswrapper[4856]: I0126 17:03:05.399188 4856 status_manager.go:851] "Failed to get status for pod" podUID="886857c0-659b-4904-b75a-c55c3f712747" pod="openshift-marketplace/redhat-marketplace-97mff" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-97mff\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:05 crc kubenswrapper[4856]: I0126 17:03:05.401723 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Jan 26 17:03:05 crc kubenswrapper[4856]: E0126 17:03:05.998167 4856 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.241:6443: connect: connection refused" interval="1.6s" Jan 26 17:03:06 crc kubenswrapper[4856]: I0126 17:03:06.229018 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lfhpz" event={"ID":"8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee","Type":"ContainerStarted","Data":"fb660ac43070a8988110315cf928def975b7819ef69dc6d82da88c39e5107bbb"} Jan 26 17:03:06 crc kubenswrapper[4856]: I0126 17:03:06.229118 4856 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:06 crc kubenswrapper[4856]: I0126 17:03:06.229308 4856 status_manager.go:851] "Failed to get status for pod" podUID="4327b726-2edc-40ad-ac96-b19a7e020048" pod="openshift-marketplace/community-operators-gdp2n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-gdp2n\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:06 crc kubenswrapper[4856]: I0126 17:03:06.229516 4856 status_manager.go:851] "Failed to get status for pod" podUID="5f34c6a8-6023-480c-a25e-46f8c4f3766b" pod="openshift-marketplace/certified-operators-bxhpt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-bxhpt\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:06 crc kubenswrapper[4856]: I0126 17:03:06.229713 4856 status_manager.go:851] "Failed to get status for pod" podUID="8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee" pod="openshift-marketplace/redhat-operators-lfhpz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-lfhpz\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:06 crc kubenswrapper[4856]: I0126 17:03:06.229861 4856 status_manager.go:851] "Failed to get status for pod" podUID="886857c0-659b-4904-b75a-c55c3f712747" pod="openshift-marketplace/redhat-marketplace-97mff" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-97mff\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:06 crc kubenswrapper[4856]: I0126 17:03:06.230005 4856 status_manager.go:851] "Failed to get status for pod" podUID="69379820-3062-4964-a8dd-8689f8cea38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:07 crc kubenswrapper[4856]: I0126 17:03:07.116591 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-97mff" Jan 26 17:03:07 crc kubenswrapper[4856]: I0126 17:03:07.117070 4856 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-97mff" Jan 26 17:03:07 crc kubenswrapper[4856]: I0126 17:03:07.157140 4856 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-97mff" Jan 26 17:03:07 crc kubenswrapper[4856]: I0126 17:03:07.157864 4856 status_manager.go:851] "Failed to get status for pod" podUID="5f34c6a8-6023-480c-a25e-46f8c4f3766b" pod="openshift-marketplace/certified-operators-bxhpt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-bxhpt\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:07 crc kubenswrapper[4856]: I0126 17:03:07.158349 4856 status_manager.go:851] "Failed to get status for pod" podUID="8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee" pod="openshift-marketplace/redhat-operators-lfhpz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-lfhpz\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:07 crc kubenswrapper[4856]: I0126 17:03:07.158791 4856 status_manager.go:851] "Failed to get status for pod" podUID="886857c0-659b-4904-b75a-c55c3f712747" pod="openshift-marketplace/redhat-marketplace-97mff" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-97mff\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:07 crc kubenswrapper[4856]: I0126 17:03:07.159088 4856 status_manager.go:851] "Failed to get status for pod" podUID="69379820-3062-4964-a8dd-8689f8cea38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:07 crc kubenswrapper[4856]: I0126 17:03:07.159373 4856 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:07 crc kubenswrapper[4856]: I0126 17:03:07.159620 4856 status_manager.go:851] "Failed to get status for pod" podUID="4327b726-2edc-40ad-ac96-b19a7e020048" pod="openshift-marketplace/community-operators-gdp2n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-gdp2n\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:07 crc kubenswrapper[4856]: I0126 17:03:07.237317 4856 generic.go:334] "Generic (PLEG): container finished" podID="8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee" containerID="fb660ac43070a8988110315cf928def975b7819ef69dc6d82da88c39e5107bbb" exitCode=0 Jan 26 17:03:07 crc kubenswrapper[4856]: I0126 17:03:07.237427 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lfhpz" event={"ID":"8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee","Type":"ContainerDied","Data":"fb660ac43070a8988110315cf928def975b7819ef69dc6d82da88c39e5107bbb"} Jan 26 17:03:07 crc kubenswrapper[4856]: I0126 17:03:07.237976 4856 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:07 crc kubenswrapper[4856]: I0126 17:03:07.238230 4856 status_manager.go:851] "Failed to get status for pod" podUID="4327b726-2edc-40ad-ac96-b19a7e020048" pod="openshift-marketplace/community-operators-gdp2n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-gdp2n\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:07 crc kubenswrapper[4856]: I0126 17:03:07.238553 4856 status_manager.go:851] "Failed to get status for pod" podUID="5f34c6a8-6023-480c-a25e-46f8c4f3766b" pod="openshift-marketplace/certified-operators-bxhpt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-bxhpt\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:07 crc kubenswrapper[4856]: I0126 17:03:07.240106 4856 status_manager.go:851] "Failed to get status for pod" podUID="8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee" pod="openshift-marketplace/redhat-operators-lfhpz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-lfhpz\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:07 crc kubenswrapper[4856]: I0126 17:03:07.240411 4856 status_manager.go:851] "Failed to get status for pod" podUID="886857c0-659b-4904-b75a-c55c3f712747" pod="openshift-marketplace/redhat-marketplace-97mff" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-97mff\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:07 crc kubenswrapper[4856]: I0126 17:03:07.240707 4856 status_manager.go:851] "Failed to get status for pod" podUID="69379820-3062-4964-a8dd-8689f8cea38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:07 crc kubenswrapper[4856]: I0126 17:03:07.276999 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-97mff" Jan 26 17:03:07 crc kubenswrapper[4856]: I0126 17:03:07.277731 4856 status_manager.go:851] "Failed to get status for pod" podUID="886857c0-659b-4904-b75a-c55c3f712747" pod="openshift-marketplace/redhat-marketplace-97mff" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-97mff\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:07 crc kubenswrapper[4856]: I0126 17:03:07.278188 4856 status_manager.go:851] "Failed to get status for pod" podUID="69379820-3062-4964-a8dd-8689f8cea38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:07 crc kubenswrapper[4856]: I0126 17:03:07.278621 4856 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:07 crc kubenswrapper[4856]: I0126 17:03:07.279010 4856 status_manager.go:851] "Failed to get status for pod" podUID="4327b726-2edc-40ad-ac96-b19a7e020048" pod="openshift-marketplace/community-operators-gdp2n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-gdp2n\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:07 crc kubenswrapper[4856]: I0126 17:03:07.279255 4856 status_manager.go:851] "Failed to get status for pod" podUID="5f34c6a8-6023-480c-a25e-46f8c4f3766b" pod="openshift-marketplace/certified-operators-bxhpt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-bxhpt\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:07 crc kubenswrapper[4856]: I0126 17:03:07.279446 4856 status_manager.go:851] "Failed to get status for pod" podUID="8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee" pod="openshift-marketplace/redhat-operators-lfhpz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-lfhpz\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:07 crc kubenswrapper[4856]: E0126 17:03:07.599666 4856 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.241:6443: connect: connection refused" interval="3.2s" Jan 26 17:03:07 crc kubenswrapper[4856]: I0126 17:03:07.810841 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gdp2n" Jan 26 17:03:07 crc kubenswrapper[4856]: I0126 17:03:07.811125 4856 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gdp2n" Jan 26 17:03:07 crc kubenswrapper[4856]: I0126 17:03:07.846286 4856 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gdp2n" Jan 26 17:03:07 crc kubenswrapper[4856]: I0126 17:03:07.846901 4856 status_manager.go:851] "Failed to get status for pod" podUID="5f34c6a8-6023-480c-a25e-46f8c4f3766b" pod="openshift-marketplace/certified-operators-bxhpt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-bxhpt\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:07 crc kubenswrapper[4856]: I0126 17:03:07.847454 4856 status_manager.go:851] "Failed to get status for pod" podUID="8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee" pod="openshift-marketplace/redhat-operators-lfhpz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-lfhpz\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:07 crc kubenswrapper[4856]: I0126 17:03:07.847657 4856 status_manager.go:851] "Failed to get status for pod" podUID="886857c0-659b-4904-b75a-c55c3f712747" pod="openshift-marketplace/redhat-marketplace-97mff" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-97mff\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:07 crc kubenswrapper[4856]: I0126 17:03:07.847812 4856 status_manager.go:851] "Failed to get status for pod" podUID="69379820-3062-4964-a8dd-8689f8cea38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:07 crc kubenswrapper[4856]: I0126 17:03:07.847976 4856 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:07 crc kubenswrapper[4856]: I0126 17:03:07.848155 4856 status_manager.go:851] "Failed to get status for pod" podUID="4327b726-2edc-40ad-ac96-b19a7e020048" pod="openshift-marketplace/community-operators-gdp2n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-gdp2n\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:08 crc kubenswrapper[4856]: I0126 17:03:08.245676 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lfhpz" event={"ID":"8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee","Type":"ContainerStarted","Data":"13caf7588d65057d336550c3ff29c21a74a680754d74d9aec3ba9f9b3471b8a6"} Jan 26 17:03:08 crc kubenswrapper[4856]: I0126 17:03:08.246057 4856 status_manager.go:851] "Failed to get status for pod" podUID="69379820-3062-4964-a8dd-8689f8cea38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:08 crc kubenswrapper[4856]: I0126 17:03:08.246328 4856 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:08 crc kubenswrapper[4856]: I0126 17:03:08.246837 4856 status_manager.go:851] "Failed to get status for pod" podUID="4327b726-2edc-40ad-ac96-b19a7e020048" pod="openshift-marketplace/community-operators-gdp2n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-gdp2n\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:08 crc kubenswrapper[4856]: I0126 17:03:08.247265 4856 status_manager.go:851] "Failed to get status for pod" podUID="5f34c6a8-6023-480c-a25e-46f8c4f3766b" pod="openshift-marketplace/certified-operators-bxhpt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-bxhpt\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:08 crc kubenswrapper[4856]: I0126 17:03:08.247455 4856 status_manager.go:851] "Failed to get status for pod" podUID="8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee" pod="openshift-marketplace/redhat-operators-lfhpz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-lfhpz\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:08 crc kubenswrapper[4856]: I0126 17:03:08.247695 4856 status_manager.go:851] "Failed to get status for pod" podUID="886857c0-659b-4904-b75a-c55c3f712747" pod="openshift-marketplace/redhat-marketplace-97mff" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-97mff\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:08 crc kubenswrapper[4856]: E0126 17:03:08.275722 4856 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/events\": dial tcp 38.102.83.241:6443: connect: connection refused" event="&Event{ObjectMeta:{redhat-marketplace-97mff.188e56a765f7220b openshift-marketplace 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-marketplace,Name:redhat-marketplace-97mff,UID:886857c0-659b-4904-b75a-c55c3f712747,APIVersion:v1,ResourceVersion:29932,FieldPath:spec.containers{registry-server},},Reason:Pulled,Message:Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\" in 441ms (442ms including waiting). Image size: 907837715 bytes.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-26 17:03:00.573405707 +0000 UTC m=+276.526659678,LastTimestamp:2026-01-26 17:03:00.573405707 +0000 UTC m=+276.526659678,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 26 17:03:08 crc kubenswrapper[4856]: I0126 17:03:08.282454 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gdp2n" Jan 26 17:03:08 crc kubenswrapper[4856]: I0126 17:03:08.283069 4856 status_manager.go:851] "Failed to get status for pod" podUID="886857c0-659b-4904-b75a-c55c3f712747" pod="openshift-marketplace/redhat-marketplace-97mff" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-97mff\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:08 crc kubenswrapper[4856]: I0126 17:03:08.283536 4856 status_manager.go:851] "Failed to get status for pod" podUID="69379820-3062-4964-a8dd-8689f8cea38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:08 crc kubenswrapper[4856]: I0126 17:03:08.283808 4856 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:08 crc kubenswrapper[4856]: I0126 17:03:08.284031 4856 status_manager.go:851] "Failed to get status for pod" podUID="4327b726-2edc-40ad-ac96-b19a7e020048" pod="openshift-marketplace/community-operators-gdp2n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-gdp2n\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:08 crc kubenswrapper[4856]: I0126 17:03:08.284273 4856 status_manager.go:851] "Failed to get status for pod" podUID="5f34c6a8-6023-480c-a25e-46f8c4f3766b" pod="openshift-marketplace/certified-operators-bxhpt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-bxhpt\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:08 crc kubenswrapper[4856]: I0126 17:03:08.284506 4856 status_manager.go:851] "Failed to get status for pod" podUID="8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee" pod="openshift-marketplace/redhat-operators-lfhpz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-lfhpz\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:09 crc kubenswrapper[4856]: I0126 17:03:09.672118 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" podUID="69008ed1-f3e5-400d-852f-adbcd94199f6" containerName="oauth-openshift" containerID="cri-o://749ef964d6b168f431c27d0286b92e40d64a8b4fb99f430b33432827ee871fc9" gracePeriod=15 Jan 26 17:03:10 crc kubenswrapper[4856]: E0126 17:03:10.800384 4856 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.241:6443: connect: connection refused" interval="6.4s" Jan 26 17:03:11 crc kubenswrapper[4856]: I0126 17:03:11.263669 4856 generic.go:334] "Generic (PLEG): container finished" podID="69008ed1-f3e5-400d-852f-adbcd94199f6" containerID="749ef964d6b168f431c27d0286b92e40d64a8b4fb99f430b33432827ee871fc9" exitCode=0 Jan 26 17:03:11 crc kubenswrapper[4856]: I0126 17:03:11.263714 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" event={"ID":"69008ed1-f3e5-400d-852f-adbcd94199f6","Type":"ContainerDied","Data":"749ef964d6b168f431c27d0286b92e40d64a8b4fb99f430b33432827ee871fc9"} Jan 26 17:03:11 crc kubenswrapper[4856]: I0126 17:03:11.971924 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:03:11 crc kubenswrapper[4856]: I0126 17:03:11.972601 4856 status_manager.go:851] "Failed to get status for pod" podUID="5f34c6a8-6023-480c-a25e-46f8c4f3766b" pod="openshift-marketplace/certified-operators-bxhpt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-bxhpt\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:11 crc kubenswrapper[4856]: I0126 17:03:11.972908 4856 status_manager.go:851] "Failed to get status for pod" podUID="8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee" pod="openshift-marketplace/redhat-operators-lfhpz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-lfhpz\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:11 crc kubenswrapper[4856]: I0126 17:03:11.973127 4856 status_manager.go:851] "Failed to get status for pod" podUID="886857c0-659b-4904-b75a-c55c3f712747" pod="openshift-marketplace/redhat-marketplace-97mff" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-97mff\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:11 crc kubenswrapper[4856]: I0126 17:03:11.973382 4856 status_manager.go:851] "Failed to get status for pod" podUID="69379820-3062-4964-a8dd-8689f8cea38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:11 crc kubenswrapper[4856]: I0126 17:03:11.973560 4856 status_manager.go:851] "Failed to get status for pod" podUID="69008ed1-f3e5-400d-852f-adbcd94199f6" pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-cb8nk\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:11 crc kubenswrapper[4856]: I0126 17:03:11.973742 4856 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:11 crc kubenswrapper[4856]: I0126 17:03:11.974090 4856 status_manager.go:851] "Failed to get status for pod" podUID="4327b726-2edc-40ad-ac96-b19a7e020048" pod="openshift-marketplace/community-operators-gdp2n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-gdp2n\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.124347 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-system-trusted-ca-bundle\") pod \"69008ed1-f3e5-400d-852f-adbcd94199f6\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.124431 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-system-serving-cert\") pod \"69008ed1-f3e5-400d-852f-adbcd94199f6\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.124463 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/69008ed1-f3e5-400d-852f-adbcd94199f6-audit-dir\") pod \"69008ed1-f3e5-400d-852f-adbcd94199f6\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.124552 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kf2b2\" (UniqueName: \"kubernetes.io/projected/69008ed1-f3e5-400d-852f-adbcd94199f6-kube-api-access-kf2b2\") pod \"69008ed1-f3e5-400d-852f-adbcd94199f6\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.124612 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-system-session\") pod \"69008ed1-f3e5-400d-852f-adbcd94199f6\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.124643 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-user-template-error\") pod \"69008ed1-f3e5-400d-852f-adbcd94199f6\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.124682 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/69008ed1-f3e5-400d-852f-adbcd94199f6-audit-policies\") pod \"69008ed1-f3e5-400d-852f-adbcd94199f6\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.124713 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-user-template-login\") pod \"69008ed1-f3e5-400d-852f-adbcd94199f6\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.124748 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-system-router-certs\") pod \"69008ed1-f3e5-400d-852f-adbcd94199f6\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.124773 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-user-idp-0-file-data\") pod \"69008ed1-f3e5-400d-852f-adbcd94199f6\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.124803 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-system-service-ca\") pod \"69008ed1-f3e5-400d-852f-adbcd94199f6\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.124873 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-system-ocp-branding-template\") pod \"69008ed1-f3e5-400d-852f-adbcd94199f6\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.124910 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-user-template-provider-selection\") pod \"69008ed1-f3e5-400d-852f-adbcd94199f6\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.124942 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-system-cliconfig\") pod \"69008ed1-f3e5-400d-852f-adbcd94199f6\" (UID: \"69008ed1-f3e5-400d-852f-adbcd94199f6\") " Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.126033 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "69008ed1-f3e5-400d-852f-adbcd94199f6" (UID: "69008ed1-f3e5-400d-852f-adbcd94199f6"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.126161 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69008ed1-f3e5-400d-852f-adbcd94199f6-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "69008ed1-f3e5-400d-852f-adbcd94199f6" (UID: "69008ed1-f3e5-400d-852f-adbcd94199f6"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.126986 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "69008ed1-f3e5-400d-852f-adbcd94199f6" (UID: "69008ed1-f3e5-400d-852f-adbcd94199f6"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.130225 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "69008ed1-f3e5-400d-852f-adbcd94199f6" (UID: "69008ed1-f3e5-400d-852f-adbcd94199f6"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.132848 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "69008ed1-f3e5-400d-852f-adbcd94199f6" (UID: "69008ed1-f3e5-400d-852f-adbcd94199f6"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.136278 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "69008ed1-f3e5-400d-852f-adbcd94199f6" (UID: "69008ed1-f3e5-400d-852f-adbcd94199f6"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.136622 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/69008ed1-f3e5-400d-852f-adbcd94199f6-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "69008ed1-f3e5-400d-852f-adbcd94199f6" (UID: "69008ed1-f3e5-400d-852f-adbcd94199f6"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.138129 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "69008ed1-f3e5-400d-852f-adbcd94199f6" (UID: "69008ed1-f3e5-400d-852f-adbcd94199f6"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.138520 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "69008ed1-f3e5-400d-852f-adbcd94199f6" (UID: "69008ed1-f3e5-400d-852f-adbcd94199f6"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.138737 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "69008ed1-f3e5-400d-852f-adbcd94199f6" (UID: "69008ed1-f3e5-400d-852f-adbcd94199f6"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.139352 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69008ed1-f3e5-400d-852f-adbcd94199f6-kube-api-access-kf2b2" (OuterVolumeSpecName: "kube-api-access-kf2b2") pod "69008ed1-f3e5-400d-852f-adbcd94199f6" (UID: "69008ed1-f3e5-400d-852f-adbcd94199f6"). InnerVolumeSpecName "kube-api-access-kf2b2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.187025 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "69008ed1-f3e5-400d-852f-adbcd94199f6" (UID: "69008ed1-f3e5-400d-852f-adbcd94199f6"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.187289 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "69008ed1-f3e5-400d-852f-adbcd94199f6" (UID: "69008ed1-f3e5-400d-852f-adbcd94199f6"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.187334 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "69008ed1-f3e5-400d-852f-adbcd94199f6" (UID: "69008ed1-f3e5-400d-852f-adbcd94199f6"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.226909 4856 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.227200 4856 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.227210 4856 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.227220 4856 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.227251 4856 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.227262 4856 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.227273 4856 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.227290 4856 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.227299 4856 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.227313 4856 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/69008ed1-f3e5-400d-852f-adbcd94199f6-audit-dir\") on node \"crc\" DevicePath \"\"" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.227323 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kf2b2\" (UniqueName: \"kubernetes.io/projected/69008ed1-f3e5-400d-852f-adbcd94199f6-kube-api-access-kf2b2\") on node \"crc\" DevicePath \"\"" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.227333 4856 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.227341 4856 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/69008ed1-f3e5-400d-852f-adbcd94199f6-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.227349 4856 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/69008ed1-f3e5-400d-852f-adbcd94199f6-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.271583 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" event={"ID":"69008ed1-f3e5-400d-852f-adbcd94199f6","Type":"ContainerDied","Data":"d2e5352f5a4f0bdf4461c4b926a9353c0b4a673c6263c30adba1a3d7a2d6a8ad"} Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.271656 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.271965 4856 scope.go:117] "RemoveContainer" containerID="749ef964d6b168f431c27d0286b92e40d64a8b4fb99f430b33432827ee871fc9" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.272604 4856 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.272929 4856 status_manager.go:851] "Failed to get status for pod" podUID="4327b726-2edc-40ad-ac96-b19a7e020048" pod="openshift-marketplace/community-operators-gdp2n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-gdp2n\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.273244 4856 status_manager.go:851] "Failed to get status for pod" podUID="5f34c6a8-6023-480c-a25e-46f8c4f3766b" pod="openshift-marketplace/certified-operators-bxhpt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-bxhpt\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.273546 4856 status_manager.go:851] "Failed to get status for pod" podUID="8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee" pod="openshift-marketplace/redhat-operators-lfhpz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-lfhpz\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.273901 4856 status_manager.go:851] "Failed to get status for pod" podUID="886857c0-659b-4904-b75a-c55c3f712747" pod="openshift-marketplace/redhat-marketplace-97mff" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-97mff\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.274242 4856 status_manager.go:851] "Failed to get status for pod" podUID="69379820-3062-4964-a8dd-8689f8cea38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.274608 4856 status_manager.go:851] "Failed to get status for pod" podUID="69008ed1-f3e5-400d-852f-adbcd94199f6" pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-cb8nk\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.295106 4856 status_manager.go:851] "Failed to get status for pod" podUID="69379820-3062-4964-a8dd-8689f8cea38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.295649 4856 status_manager.go:851] "Failed to get status for pod" podUID="69008ed1-f3e5-400d-852f-adbcd94199f6" pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-cb8nk\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.295955 4856 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.296230 4856 status_manager.go:851] "Failed to get status for pod" podUID="4327b726-2edc-40ad-ac96-b19a7e020048" pod="openshift-marketplace/community-operators-gdp2n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-gdp2n\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.296564 4856 status_manager.go:851] "Failed to get status for pod" podUID="5f34c6a8-6023-480c-a25e-46f8c4f3766b" pod="openshift-marketplace/certified-operators-bxhpt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-bxhpt\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.296821 4856 status_manager.go:851] "Failed to get status for pod" podUID="8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee" pod="openshift-marketplace/redhat-operators-lfhpz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-lfhpz\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:12 crc kubenswrapper[4856]: I0126 17:03:12.297058 4856 status_manager.go:851] "Failed to get status for pod" podUID="886857c0-659b-4904-b75a-c55c3f712747" pod="openshift-marketplace/redhat-marketplace-97mff" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-97mff\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:14 crc kubenswrapper[4856]: I0126 17:03:14.394946 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 17:03:14 crc kubenswrapper[4856]: I0126 17:03:14.395869 4856 status_manager.go:851] "Failed to get status for pod" podUID="5f34c6a8-6023-480c-a25e-46f8c4f3766b" pod="openshift-marketplace/certified-operators-bxhpt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-bxhpt\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:14 crc kubenswrapper[4856]: I0126 17:03:14.396309 4856 status_manager.go:851] "Failed to get status for pod" podUID="8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee" pod="openshift-marketplace/redhat-operators-lfhpz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-lfhpz\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:14 crc kubenswrapper[4856]: I0126 17:03:14.396902 4856 status_manager.go:851] "Failed to get status for pod" podUID="886857c0-659b-4904-b75a-c55c3f712747" pod="openshift-marketplace/redhat-marketplace-97mff" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-97mff\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:14 crc kubenswrapper[4856]: I0126 17:03:14.397211 4856 status_manager.go:851] "Failed to get status for pod" podUID="69379820-3062-4964-a8dd-8689f8cea38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:14 crc kubenswrapper[4856]: I0126 17:03:14.397669 4856 status_manager.go:851] "Failed to get status for pod" podUID="69008ed1-f3e5-400d-852f-adbcd94199f6" pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-cb8nk\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:14 crc kubenswrapper[4856]: I0126 17:03:14.397969 4856 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:14 crc kubenswrapper[4856]: I0126 17:03:14.398214 4856 status_manager.go:851] "Failed to get status for pod" podUID="4327b726-2edc-40ad-ac96-b19a7e020048" pod="openshift-marketplace/community-operators-gdp2n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-gdp2n\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:14 crc kubenswrapper[4856]: I0126 17:03:14.410273 4856 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="59ecd87a-c5db-446d-ad3e-cfabbd648c1d" Jan 26 17:03:14 crc kubenswrapper[4856]: I0126 17:03:14.410330 4856 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="59ecd87a-c5db-446d-ad3e-cfabbd648c1d" Jan 26 17:03:14 crc kubenswrapper[4856]: E0126 17:03:14.410881 4856 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 17:03:14 crc kubenswrapper[4856]: I0126 17:03:14.443518 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.311083 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.311402 4856 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="fb3c5348b8b83991cbb42255dc07d74fe50e200793efe1a7b2b2727a5c2be800" exitCode=1 Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.311498 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"fb3c5348b8b83991cbb42255dc07d74fe50e200793efe1a7b2b2727a5c2be800"} Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.312064 4856 scope.go:117] "RemoveContainer" containerID="fb3c5348b8b83991cbb42255dc07d74fe50e200793efe1a7b2b2727a5c2be800" Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.312482 4856 status_manager.go:851] "Failed to get status for pod" podUID="69379820-3062-4964-a8dd-8689f8cea38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.312741 4856 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lfhpz" Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.312799 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lfhpz" Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.312973 4856 status_manager.go:851] "Failed to get status for pod" podUID="69008ed1-f3e5-400d-852f-adbcd94199f6" pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-cb8nk\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.313848 4856 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.323954 4856 status_manager.go:851] "Failed to get status for pod" podUID="4327b726-2edc-40ad-ac96-b19a7e020048" pod="openshift-marketplace/community-operators-gdp2n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-gdp2n\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.324997 4856 status_manager.go:851] "Failed to get status for pod" podUID="5f34c6a8-6023-480c-a25e-46f8c4f3766b" pod="openshift-marketplace/certified-operators-bxhpt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-bxhpt\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.325590 4856 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="0f264a73611fd2a7e43252672774079b6705528a6b5a493040487d1b27e3dc7e" exitCode=0 Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.325639 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"0f264a73611fd2a7e43252672774079b6705528a6b5a493040487d1b27e3dc7e"} Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.325711 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b454fc4809f34d7d4bf4b768d40eaf90ad0aebd9fa27c08031ff2c9cfd3e6b1e"} Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.325774 4856 status_manager.go:851] "Failed to get status for pod" podUID="8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee" pod="openshift-marketplace/redhat-operators-lfhpz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-lfhpz\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.326115 4856 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="59ecd87a-c5db-446d-ad3e-cfabbd648c1d" Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.326134 4856 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="59ecd87a-c5db-446d-ad3e-cfabbd648c1d" Jan 26 17:03:15 crc kubenswrapper[4856]: E0126 17:03:15.326483 4856 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.326769 4856 status_manager.go:851] "Failed to get status for pod" podUID="886857c0-659b-4904-b75a-c55c3f712747" pod="openshift-marketplace/redhat-marketplace-97mff" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-97mff\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.327662 4856 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.329727 4856 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.330210 4856 status_manager.go:851] "Failed to get status for pod" podUID="4327b726-2edc-40ad-ac96-b19a7e020048" pod="openshift-marketplace/community-operators-gdp2n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-gdp2n\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.331980 4856 status_manager.go:851] "Failed to get status for pod" podUID="5f34c6a8-6023-480c-a25e-46f8c4f3766b" pod="openshift-marketplace/certified-operators-bxhpt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-bxhpt\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.332508 4856 status_manager.go:851] "Failed to get status for pod" podUID="8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee" pod="openshift-marketplace/redhat-operators-lfhpz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-lfhpz\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.333323 4856 status_manager.go:851] "Failed to get status for pod" podUID="886857c0-659b-4904-b75a-c55c3f712747" pod="openshift-marketplace/redhat-marketplace-97mff" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-97mff\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.334174 4856 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.334621 4856 status_manager.go:851] "Failed to get status for pod" podUID="69379820-3062-4964-a8dd-8689f8cea38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.335064 4856 status_manager.go:851] "Failed to get status for pod" podUID="69008ed1-f3e5-400d-852f-adbcd94199f6" pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-cb8nk\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.360894 4856 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lfhpz" Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.361432 4856 status_manager.go:851] "Failed to get status for pod" podUID="69008ed1-f3e5-400d-852f-adbcd94199f6" pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-cb8nk\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.361816 4856 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.362996 4856 status_manager.go:851] "Failed to get status for pod" podUID="4327b726-2edc-40ad-ac96-b19a7e020048" pod="openshift-marketplace/community-operators-gdp2n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-gdp2n\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.363164 4856 status_manager.go:851] "Failed to get status for pod" podUID="5f34c6a8-6023-480c-a25e-46f8c4f3766b" pod="openshift-marketplace/certified-operators-bxhpt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-bxhpt\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.363308 4856 status_manager.go:851] "Failed to get status for pod" podUID="8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee" pod="openshift-marketplace/redhat-operators-lfhpz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-lfhpz\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.363453 4856 status_manager.go:851] "Failed to get status for pod" podUID="886857c0-659b-4904-b75a-c55c3f712747" pod="openshift-marketplace/redhat-marketplace-97mff" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-97mff\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.363957 4856 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.365171 4856 status_manager.go:851] "Failed to get status for pod" podUID="69379820-3062-4964-a8dd-8689f8cea38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.442912 4856 status_manager.go:851] "Failed to get status for pod" podUID="69008ed1-f3e5-400d-852f-adbcd94199f6" pod="openshift-authentication/oauth-openshift-558db77b4-cb8nk" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-cb8nk\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.443463 4856 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.443639 4856 status_manager.go:851] "Failed to get status for pod" podUID="4327b726-2edc-40ad-ac96-b19a7e020048" pod="openshift-marketplace/community-operators-gdp2n" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-gdp2n\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.443793 4856 status_manager.go:851] "Failed to get status for pod" podUID="5f34c6a8-6023-480c-a25e-46f8c4f3766b" pod="openshift-marketplace/certified-operators-bxhpt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-bxhpt\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.443987 4856 status_manager.go:851] "Failed to get status for pod" podUID="8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee" pod="openshift-marketplace/redhat-operators-lfhpz" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-lfhpz\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.444163 4856 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.444414 4856 status_manager.go:851] "Failed to get status for pod" podUID="886857c0-659b-4904-b75a-c55c3f712747" pod="openshift-marketplace/redhat-marketplace-97mff" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-97mff\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.444637 4856 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:15 crc kubenswrapper[4856]: I0126 17:03:15.444797 4856 status_manager.go:851] "Failed to get status for pod" podUID="69379820-3062-4964-a8dd-8689f8cea38d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:15 crc kubenswrapper[4856]: E0126 17:03:15.623611 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T17:03:15Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T17:03:15Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T17:03:15Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T17:03:15Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:15 crc kubenswrapper[4856]: E0126 17:03:15.623912 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:15 crc kubenswrapper[4856]: E0126 17:03:15.624177 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:15 crc kubenswrapper[4856]: E0126 17:03:15.624398 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:15 crc kubenswrapper[4856]: E0126 17:03:15.624785 4856 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.241:6443: connect: connection refused" Jan 26 17:03:15 crc kubenswrapper[4856]: E0126 17:03:15.624816 4856 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 26 17:03:16 crc kubenswrapper[4856]: I0126 17:03:16.335203 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 26 17:03:16 crc kubenswrapper[4856]: I0126 17:03:16.335331 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"80273e46ebcca0298cf3f63e0e0aabceb330a19fc9f5399a09ac60d75bf71e10"} Jan 26 17:03:16 crc kubenswrapper[4856]: I0126 17:03:16.337621 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b2d6eca8f4d2929c26b180b5010215bd6e8b2a977125d6a5f2c070fabb3ddee8"} Jan 26 17:03:16 crc kubenswrapper[4856]: I0126 17:03:16.337670 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"498cf9d3ffe151aa5744d939a0d77cf3881afe48bae0b6e4620ce96a1cd0014c"} Jan 26 17:03:16 crc kubenswrapper[4856]: I0126 17:03:16.391809 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lfhpz" Jan 26 17:03:17 crc kubenswrapper[4856]: I0126 17:03:17.345606 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"8cdb077ef075ef70025deaf712b29a0b97f686ddd45a68c8741b2541bb6a5fad"} Jan 26 17:03:19 crc kubenswrapper[4856]: I0126 17:03:19.282392 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" podUID="cfa40861-cc08-4145-a185-6a3fb07eaabe" containerName="registry" containerID="cri-o://fc8e05e1e87fe66232302aff71c23d6b6c36b366751f113f41815a46bc948eb9" gracePeriod=30 Jan 26 17:03:19 crc kubenswrapper[4856]: I0126 17:03:19.361418 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"aa8e49724230f0e647de2fe59a2cd9370bd08515b0927af18b69b392f4dfd64b"} Jan 26 17:03:19 crc kubenswrapper[4856]: I0126 17:03:19.423491 4856 patch_prober.go:28] interesting pod/image-registry-697d97f7c8-wxbdh container/registry namespace/openshift-image-registry: Readiness probe status=failure output="Get \"https://10.217.0.30:5000/healthz\": dial tcp 10.217.0.30:5000: connect: connection refused" start-of-body= Jan 26 17:03:19 crc kubenswrapper[4856]: I0126 17:03:19.423608 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" podUID="cfa40861-cc08-4145-a185-6a3fb07eaabe" containerName="registry" probeResult="failure" output="Get \"https://10.217.0.30:5000/healthz\": dial tcp 10.217.0.30:5000: connect: connection refused" Jan 26 17:03:20 crc kubenswrapper[4856]: I0126 17:03:20.368868 4856 generic.go:334] "Generic (PLEG): container finished" podID="cfa40861-cc08-4145-a185-6a3fb07eaabe" containerID="fc8e05e1e87fe66232302aff71c23d6b6c36b366751f113f41815a46bc948eb9" exitCode=0 Jan 26 17:03:20 crc kubenswrapper[4856]: I0126 17:03:20.369003 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" event={"ID":"cfa40861-cc08-4145-a185-6a3fb07eaabe","Type":"ContainerDied","Data":"fc8e05e1e87fe66232302aff71c23d6b6c36b366751f113f41815a46bc948eb9"} Jan 26 17:03:20 crc kubenswrapper[4856]: I0126 17:03:20.373114 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"0c3a9c85a068db8311cb2e29673d265412a4ac44329f3e596c834ede0310716f"} Jan 26 17:03:20 crc kubenswrapper[4856]: I0126 17:03:20.373337 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 17:03:20 crc kubenswrapper[4856]: I0126 17:03:20.373452 4856 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="59ecd87a-c5db-446d-ad3e-cfabbd648c1d" Jan 26 17:03:20 crc kubenswrapper[4856]: I0126 17:03:20.373475 4856 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="59ecd87a-c5db-446d-ad3e-cfabbd648c1d" Jan 26 17:03:20 crc kubenswrapper[4856]: I0126 17:03:20.382386 4856 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 17:03:20 crc kubenswrapper[4856]: I0126 17:03:20.583904 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 17:03:20 crc kubenswrapper[4856]: I0126 17:03:20.891575 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:03:20 crc kubenswrapper[4856]: I0126 17:03:20.946826 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cfa40861-cc08-4145-a185-6a3fb07eaabe-bound-sa-token\") pod \"cfa40861-cc08-4145-a185-6a3fb07eaabe\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " Jan 26 17:03:20 crc kubenswrapper[4856]: I0126 17:03:20.946875 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cfa40861-cc08-4145-a185-6a3fb07eaabe-trusted-ca\") pod \"cfa40861-cc08-4145-a185-6a3fb07eaabe\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " Jan 26 17:03:20 crc kubenswrapper[4856]: I0126 17:03:20.946897 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cfa40861-cc08-4145-a185-6a3fb07eaabe-ca-trust-extracted\") pod \"cfa40861-cc08-4145-a185-6a3fb07eaabe\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " Jan 26 17:03:20 crc kubenswrapper[4856]: I0126 17:03:20.946939 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tf448\" (UniqueName: \"kubernetes.io/projected/cfa40861-cc08-4145-a185-6a3fb07eaabe-kube-api-access-tf448\") pod \"cfa40861-cc08-4145-a185-6a3fb07eaabe\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " Jan 26 17:03:20 crc kubenswrapper[4856]: I0126 17:03:20.947145 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"cfa40861-cc08-4145-a185-6a3fb07eaabe\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " Jan 26 17:03:20 crc kubenswrapper[4856]: I0126 17:03:20.947215 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cfa40861-cc08-4145-a185-6a3fb07eaabe-installation-pull-secrets\") pod \"cfa40861-cc08-4145-a185-6a3fb07eaabe\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " Jan 26 17:03:20 crc kubenswrapper[4856]: I0126 17:03:20.947236 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cfa40861-cc08-4145-a185-6a3fb07eaabe-registry-tls\") pod \"cfa40861-cc08-4145-a185-6a3fb07eaabe\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " Jan 26 17:03:20 crc kubenswrapper[4856]: I0126 17:03:20.947254 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cfa40861-cc08-4145-a185-6a3fb07eaabe-registry-certificates\") pod \"cfa40861-cc08-4145-a185-6a3fb07eaabe\" (UID: \"cfa40861-cc08-4145-a185-6a3fb07eaabe\") " Jan 26 17:03:20 crc kubenswrapper[4856]: I0126 17:03:20.947783 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cfa40861-cc08-4145-a185-6a3fb07eaabe-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "cfa40861-cc08-4145-a185-6a3fb07eaabe" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:03:20 crc kubenswrapper[4856]: I0126 17:03:20.948162 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cfa40861-cc08-4145-a185-6a3fb07eaabe-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "cfa40861-cc08-4145-a185-6a3fb07eaabe" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:03:20 crc kubenswrapper[4856]: I0126 17:03:20.952851 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfa40861-cc08-4145-a185-6a3fb07eaabe-kube-api-access-tf448" (OuterVolumeSpecName: "kube-api-access-tf448") pod "cfa40861-cc08-4145-a185-6a3fb07eaabe" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe"). InnerVolumeSpecName "kube-api-access-tf448". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:03:20 crc kubenswrapper[4856]: I0126 17:03:20.953080 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfa40861-cc08-4145-a185-6a3fb07eaabe-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "cfa40861-cc08-4145-a185-6a3fb07eaabe" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:03:20 crc kubenswrapper[4856]: I0126 17:03:20.953299 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfa40861-cc08-4145-a185-6a3fb07eaabe-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "cfa40861-cc08-4145-a185-6a3fb07eaabe" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:03:20 crc kubenswrapper[4856]: I0126 17:03:20.953445 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfa40861-cc08-4145-a185-6a3fb07eaabe-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "cfa40861-cc08-4145-a185-6a3fb07eaabe" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:03:20 crc kubenswrapper[4856]: I0126 17:03:20.956725 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "cfa40861-cc08-4145-a185-6a3fb07eaabe" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 26 17:03:20 crc kubenswrapper[4856]: I0126 17:03:20.982714 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cfa40861-cc08-4145-a185-6a3fb07eaabe-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "cfa40861-cc08-4145-a185-6a3fb07eaabe" (UID: "cfa40861-cc08-4145-a185-6a3fb07eaabe"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:03:21 crc kubenswrapper[4856]: I0126 17:03:21.048838 4856 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/cfa40861-cc08-4145-a185-6a3fb07eaabe-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Jan 26 17:03:21 crc kubenswrapper[4856]: I0126 17:03:21.048885 4856 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/cfa40861-cc08-4145-a185-6a3fb07eaabe-registry-tls\") on node \"crc\" DevicePath \"\"" Jan 26 17:03:21 crc kubenswrapper[4856]: I0126 17:03:21.048900 4856 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/cfa40861-cc08-4145-a185-6a3fb07eaabe-registry-certificates\") on node \"crc\" DevicePath \"\"" Jan 26 17:03:21 crc kubenswrapper[4856]: I0126 17:03:21.048911 4856 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/cfa40861-cc08-4145-a185-6a3fb07eaabe-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 26 17:03:21 crc kubenswrapper[4856]: I0126 17:03:21.048923 4856 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cfa40861-cc08-4145-a185-6a3fb07eaabe-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 26 17:03:21 crc kubenswrapper[4856]: I0126 17:03:21.048935 4856 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/cfa40861-cc08-4145-a185-6a3fb07eaabe-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Jan 26 17:03:21 crc kubenswrapper[4856]: I0126 17:03:21.048946 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tf448\" (UniqueName: \"kubernetes.io/projected/cfa40861-cc08-4145-a185-6a3fb07eaabe-kube-api-access-tf448\") on node \"crc\" DevicePath \"\"" Jan 26 17:03:21 crc kubenswrapper[4856]: I0126 17:03:21.380048 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" event={"ID":"cfa40861-cc08-4145-a185-6a3fb07eaabe","Type":"ContainerDied","Data":"ae7df2de181ac684cadd8c52c3b8878c72703f16549d24e92a2fc45b186ce717"} Jan 26 17:03:21 crc kubenswrapper[4856]: I0126 17:03:21.380062 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-wxbdh" Jan 26 17:03:21 crc kubenswrapper[4856]: I0126 17:03:21.380107 4856 scope.go:117] "RemoveContainer" containerID="fc8e05e1e87fe66232302aff71c23d6b6c36b366751f113f41815a46bc948eb9" Jan 26 17:03:21 crc kubenswrapper[4856]: I0126 17:03:21.380352 4856 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="59ecd87a-c5db-446d-ad3e-cfabbd648c1d" Jan 26 17:03:21 crc kubenswrapper[4856]: I0126 17:03:21.380369 4856 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="59ecd87a-c5db-446d-ad3e-cfabbd648c1d" Jan 26 17:03:22 crc kubenswrapper[4856]: I0126 17:03:22.879606 4856 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="1c355228-5a24-45a2-9876-b4f0732a65d0" Jan 26 17:03:23 crc kubenswrapper[4856]: I0126 17:03:23.508739 4856 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 17:03:23 crc kubenswrapper[4856]: I0126 17:03:23.515290 4856 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 17:03:24 crc kubenswrapper[4856]: I0126 17:03:24.939304 4856 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Jan 26 17:03:30 crc kubenswrapper[4856]: I0126 17:03:30.589298 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 17:03:32 crc kubenswrapper[4856]: I0126 17:03:32.758903 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Jan 26 17:03:32 crc kubenswrapper[4856]: I0126 17:03:32.844037 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 26 17:03:33 crc kubenswrapper[4856]: I0126 17:03:33.402891 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Jan 26 17:03:33 crc kubenswrapper[4856]: I0126 17:03:33.854541 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Jan 26 17:03:33 crc kubenswrapper[4856]: I0126 17:03:33.901157 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Jan 26 17:03:33 crc kubenswrapper[4856]: I0126 17:03:33.995367 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Jan 26 17:03:34 crc kubenswrapper[4856]: I0126 17:03:34.278793 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Jan 26 17:03:34 crc kubenswrapper[4856]: I0126 17:03:34.530747 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Jan 26 17:03:34 crc kubenswrapper[4856]: I0126 17:03:34.539666 4856 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","pod69379820-3062-4964-a8dd-8689f8cea38d"] err="unable to destroy cgroup paths for cgroup [kubepods pod69379820-3062-4964-a8dd-8689f8cea38d] : Timed out while waiting for systemd to remove kubepods-pod69379820_3062_4964_a8dd_8689f8cea38d.slice" Jan 26 17:03:34 crc kubenswrapper[4856]: E0126 17:03:34.539998 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods pod69379820-3062-4964-a8dd-8689f8cea38d] : unable to destroy cgroup paths for cgroup [kubepods pod69379820-3062-4964-a8dd-8689f8cea38d] : Timed out while waiting for systemd to remove kubepods-pod69379820_3062_4964_a8dd_8689f8cea38d.slice" pod="openshift-kube-apiserver/installer-9-crc" podUID="69379820-3062-4964-a8dd-8689f8cea38d" Jan 26 17:03:34 crc kubenswrapper[4856]: I0126 17:03:34.647628 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Jan 26 17:03:34 crc kubenswrapper[4856]: I0126 17:03:34.805366 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Jan 26 17:03:34 crc kubenswrapper[4856]: I0126 17:03:34.809673 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Jan 26 17:03:34 crc kubenswrapper[4856]: I0126 17:03:34.854428 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Jan 26 17:03:34 crc kubenswrapper[4856]: I0126 17:03:34.894005 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Jan 26 17:03:35 crc kubenswrapper[4856]: I0126 17:03:35.004192 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Jan 26 17:03:35 crc kubenswrapper[4856]: I0126 17:03:35.035288 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 26 17:03:35 crc kubenswrapper[4856]: I0126 17:03:35.202142 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Jan 26 17:03:35 crc kubenswrapper[4856]: I0126 17:03:35.328365 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 26 17:03:35 crc kubenswrapper[4856]: I0126 17:03:35.526370 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 26 17:03:35 crc kubenswrapper[4856]: I0126 17:03:35.593299 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Jan 26 17:03:35 crc kubenswrapper[4856]: I0126 17:03:35.637103 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Jan 26 17:03:35 crc kubenswrapper[4856]: I0126 17:03:35.766985 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Jan 26 17:03:35 crc kubenswrapper[4856]: I0126 17:03:35.997628 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Jan 26 17:03:36 crc kubenswrapper[4856]: I0126 17:03:36.027705 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Jan 26 17:03:36 crc kubenswrapper[4856]: I0126 17:03:36.031989 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Jan 26 17:03:36 crc kubenswrapper[4856]: I0126 17:03:36.129043 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Jan 26 17:03:36 crc kubenswrapper[4856]: I0126 17:03:36.170563 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Jan 26 17:03:36 crc kubenswrapper[4856]: I0126 17:03:36.171976 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Jan 26 17:03:36 crc kubenswrapper[4856]: I0126 17:03:36.259288 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Jan 26 17:03:36 crc kubenswrapper[4856]: I0126 17:03:36.527088 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Jan 26 17:03:36 crc kubenswrapper[4856]: I0126 17:03:36.545665 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Jan 26 17:03:36 crc kubenswrapper[4856]: I0126 17:03:36.550380 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Jan 26 17:03:36 crc kubenswrapper[4856]: I0126 17:03:36.630298 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Jan 26 17:03:36 crc kubenswrapper[4856]: I0126 17:03:36.647645 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Jan 26 17:03:36 crc kubenswrapper[4856]: I0126 17:03:36.654972 4856 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Jan 26 17:03:36 crc kubenswrapper[4856]: I0126 17:03:36.810107 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Jan 26 17:03:36 crc kubenswrapper[4856]: I0126 17:03:36.836259 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Jan 26 17:03:36 crc kubenswrapper[4856]: I0126 17:03:36.866455 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Jan 26 17:03:36 crc kubenswrapper[4856]: I0126 17:03:36.967431 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Jan 26 17:03:37 crc kubenswrapper[4856]: I0126 17:03:37.044551 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Jan 26 17:03:37 crc kubenswrapper[4856]: I0126 17:03:37.143128 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Jan 26 17:03:37 crc kubenswrapper[4856]: I0126 17:03:37.172434 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Jan 26 17:03:37 crc kubenswrapper[4856]: I0126 17:03:37.192564 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Jan 26 17:03:37 crc kubenswrapper[4856]: I0126 17:03:37.200011 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Jan 26 17:03:37 crc kubenswrapper[4856]: I0126 17:03:37.254171 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Jan 26 17:03:37 crc kubenswrapper[4856]: I0126 17:03:37.459294 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Jan 26 17:03:37 crc kubenswrapper[4856]: I0126 17:03:37.523615 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Jan 26 17:03:37 crc kubenswrapper[4856]: I0126 17:03:37.548772 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Jan 26 17:03:37 crc kubenswrapper[4856]: I0126 17:03:37.563986 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Jan 26 17:03:37 crc kubenswrapper[4856]: I0126 17:03:37.567956 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Jan 26 17:03:37 crc kubenswrapper[4856]: I0126 17:03:37.570732 4856 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Jan 26 17:03:37 crc kubenswrapper[4856]: I0126 17:03:37.591192 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Jan 26 17:03:37 crc kubenswrapper[4856]: I0126 17:03:37.675782 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 26 17:03:37 crc kubenswrapper[4856]: I0126 17:03:37.692820 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Jan 26 17:03:37 crc kubenswrapper[4856]: I0126 17:03:37.709070 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Jan 26 17:03:37 crc kubenswrapper[4856]: I0126 17:03:37.712369 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Jan 26 17:03:37 crc kubenswrapper[4856]: I0126 17:03:37.764482 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Jan 26 17:03:37 crc kubenswrapper[4856]: I0126 17:03:37.816619 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Jan 26 17:03:37 crc kubenswrapper[4856]: I0126 17:03:37.846995 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Jan 26 17:03:37 crc kubenswrapper[4856]: I0126 17:03:37.908759 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Jan 26 17:03:37 crc kubenswrapper[4856]: I0126 17:03:37.945172 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Jan 26 17:03:38 crc kubenswrapper[4856]: I0126 17:03:38.026161 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Jan 26 17:03:38 crc kubenswrapper[4856]: I0126 17:03:38.055222 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 26 17:03:38 crc kubenswrapper[4856]: I0126 17:03:38.066228 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Jan 26 17:03:38 crc kubenswrapper[4856]: I0126 17:03:38.074446 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Jan 26 17:03:38 crc kubenswrapper[4856]: I0126 17:03:38.254241 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Jan 26 17:03:38 crc kubenswrapper[4856]: I0126 17:03:38.270790 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Jan 26 17:03:38 crc kubenswrapper[4856]: I0126 17:03:38.346569 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 26 17:03:38 crc kubenswrapper[4856]: I0126 17:03:38.360031 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Jan 26 17:03:38 crc kubenswrapper[4856]: I0126 17:03:38.434095 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Jan 26 17:03:38 crc kubenswrapper[4856]: I0126 17:03:38.466431 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Jan 26 17:03:38 crc kubenswrapper[4856]: I0126 17:03:38.468738 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Jan 26 17:03:38 crc kubenswrapper[4856]: I0126 17:03:38.508989 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Jan 26 17:03:38 crc kubenswrapper[4856]: I0126 17:03:38.539174 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Jan 26 17:03:38 crc kubenswrapper[4856]: I0126 17:03:38.544340 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Jan 26 17:03:38 crc kubenswrapper[4856]: I0126 17:03:38.617969 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Jan 26 17:03:38 crc kubenswrapper[4856]: I0126 17:03:38.630632 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Jan 26 17:03:38 crc kubenswrapper[4856]: I0126 17:03:38.831607 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Jan 26 17:03:38 crc kubenswrapper[4856]: I0126 17:03:38.832326 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Jan 26 17:03:38 crc kubenswrapper[4856]: I0126 17:03:38.965764 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Jan 26 17:03:38 crc kubenswrapper[4856]: I0126 17:03:38.987698 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Jan 26 17:03:38 crc kubenswrapper[4856]: I0126 17:03:38.988432 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.087404 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.130282 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.171662 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.175385 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.389581 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.474960 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.538428 4856 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.538802 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-97mff" podStartSLOduration=41.079029315 podStartE2EDuration="43.538785079s" podCreationTimestamp="2026-01-26 17:02:56 +0000 UTC" firstStartedPulling="2026-01-26 17:02:58.113635403 +0000 UTC m=+274.066889384" lastFinishedPulling="2026-01-26 17:03:00.573391167 +0000 UTC m=+276.526645148" observedRunningTime="2026-01-26 17:03:22.893187657 +0000 UTC m=+298.846441648" watchObservedRunningTime="2026-01-26 17:03:39.538785079 +0000 UTC m=+315.492039060" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.539677 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lfhpz" podStartSLOduration=33.877414329 podStartE2EDuration="45.539673774s" podCreationTimestamp="2026-01-26 17:02:54 +0000 UTC" firstStartedPulling="2026-01-26 17:02:55.972045452 +0000 UTC m=+271.925299433" lastFinishedPulling="2026-01-26 17:03:07.634304897 +0000 UTC m=+283.587558878" observedRunningTime="2026-01-26 17:03:22.876729534 +0000 UTC m=+298.829983525" watchObservedRunningTime="2026-01-26 17:03:39.539673774 +0000 UTC m=+315.492927755" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.540788 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=39.540784266 podStartE2EDuration="39.540784266s" podCreationTimestamp="2026-01-26 17:03:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:03:22.960055926 +0000 UTC m=+298.913309907" watchObservedRunningTime="2026-01-26 17:03:39.540784266 +0000 UTC m=+315.494038247" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.540997 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gdp2n" podStartSLOduration=40.073558069 podStartE2EDuration="42.540992512s" podCreationTimestamp="2026-01-26 17:02:57 +0000 UTC" firstStartedPulling="2026-01-26 17:02:59.121501492 +0000 UTC m=+275.074755473" lastFinishedPulling="2026-01-26 17:03:01.588935945 +0000 UTC m=+277.542189916" observedRunningTime="2026-01-26 17:03:22.843797389 +0000 UTC m=+298.797051400" watchObservedRunningTime="2026-01-26 17:03:39.540992512 +0000 UTC m=+315.494246493" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.542299 4856 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-cb8nk","openshift-image-registry/image-registry-697d97f7c8-wxbdh","openshift-kube-apiserver/kube-apiserver-crc"] Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.542348 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-5d4455474f-tnwpf"] Jan 26 17:03:39 crc kubenswrapper[4856]: E0126 17:03:39.542563 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69379820-3062-4964-a8dd-8689f8cea38d" containerName="installer" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.542581 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="69379820-3062-4964-a8dd-8689f8cea38d" containerName="installer" Jan 26 17:03:39 crc kubenswrapper[4856]: E0126 17:03:39.542592 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69008ed1-f3e5-400d-852f-adbcd94199f6" containerName="oauth-openshift" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.542598 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="69008ed1-f3e5-400d-852f-adbcd94199f6" containerName="oauth-openshift" Jan 26 17:03:39 crc kubenswrapper[4856]: E0126 17:03:39.542608 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfa40861-cc08-4145-a185-6a3fb07eaabe" containerName="registry" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.542614 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfa40861-cc08-4145-a185-6a3fb07eaabe" containerName="registry" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.542723 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="69379820-3062-4964-a8dd-8689f8cea38d" containerName="installer" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.542733 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="69008ed1-f3e5-400d-852f-adbcd94199f6" containerName="oauth-openshift" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.542746 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfa40861-cc08-4145-a185-6a3fb07eaabe" containerName="registry" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.543108 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.546574 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.547069 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.547074 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.547619 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.547873 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.547876 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.548220 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.548408 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.549846 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.549863 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.549979 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.551010 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.551220 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.560498 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.568218 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.588665 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.588696 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=19.58866649 podStartE2EDuration="19.58866649s" podCreationTimestamp="2026-01-26 17:03:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:03:39.58132612 +0000 UTC m=+315.534580111" watchObservedRunningTime="2026-01-26 17:03:39.58866649 +0000 UTC m=+315.541920491" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.591564 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.651364 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a-v4-0-config-system-session\") pod \"oauth-openshift-5d4455474f-tnwpf\" (UID: \"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a\") " pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.651861 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a-v4-0-config-user-template-login\") pod \"oauth-openshift-5d4455474f-tnwpf\" (UID: \"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a\") " pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.652106 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5d4455474f-tnwpf\" (UID: \"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a\") " pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.652295 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a-audit-policies\") pod \"oauth-openshift-5d4455474f-tnwpf\" (UID: \"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a\") " pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.652641 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5d4455474f-tnwpf\" (UID: \"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a\") " pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.652811 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a-v4-0-config-system-router-certs\") pod \"oauth-openshift-5d4455474f-tnwpf\" (UID: \"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a\") " pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.652936 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a-audit-dir\") pod \"oauth-openshift-5d4455474f-tnwpf\" (UID: \"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a\") " pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.653075 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a-v4-0-config-user-template-error\") pod \"oauth-openshift-5d4455474f-tnwpf\" (UID: \"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a\") " pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.653220 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a-v4-0-config-system-service-ca\") pod \"oauth-openshift-5d4455474f-tnwpf\" (UID: \"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a\") " pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.653338 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5d4455474f-tnwpf\" (UID: \"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a\") " pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.653775 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5d4455474f-tnwpf\" (UID: \"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a\") " pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.654086 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5d4455474f-tnwpf\" (UID: \"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a\") " pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.654134 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5d4455474f-tnwpf\" (UID: \"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a\") " pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.654286 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-956kp\" (UniqueName: \"kubernetes.io/projected/34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a-kube-api-access-956kp\") pod \"oauth-openshift-5d4455474f-tnwpf\" (UID: \"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a\") " pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.742592 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.755423 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5d4455474f-tnwpf\" (UID: \"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a\") " pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.755747 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a-v4-0-config-system-router-certs\") pod \"oauth-openshift-5d4455474f-tnwpf\" (UID: \"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a\") " pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.755865 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a-audit-dir\") pod \"oauth-openshift-5d4455474f-tnwpf\" (UID: \"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a\") " pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.755955 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a-v4-0-config-user-template-error\") pod \"oauth-openshift-5d4455474f-tnwpf\" (UID: \"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a\") " pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.756041 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a-v4-0-config-system-service-ca\") pod \"oauth-openshift-5d4455474f-tnwpf\" (UID: \"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a\") " pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.756119 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5d4455474f-tnwpf\" (UID: \"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a\") " pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.756241 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5d4455474f-tnwpf\" (UID: \"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a\") " pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.756323 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5d4455474f-tnwpf\" (UID: \"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a\") " pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.756428 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5d4455474f-tnwpf\" (UID: \"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a\") " pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.756543 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-956kp\" (UniqueName: \"kubernetes.io/projected/34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a-kube-api-access-956kp\") pod \"oauth-openshift-5d4455474f-tnwpf\" (UID: \"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a\") " pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.756632 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a-v4-0-config-system-session\") pod \"oauth-openshift-5d4455474f-tnwpf\" (UID: \"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a\") " pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.756724 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a-v4-0-config-user-template-login\") pod \"oauth-openshift-5d4455474f-tnwpf\" (UID: \"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a\") " pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.756826 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5d4455474f-tnwpf\" (UID: \"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a\") " pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.756910 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a-audit-policies\") pod \"oauth-openshift-5d4455474f-tnwpf\" (UID: \"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a\") " pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.757461 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a-v4-0-config-system-service-ca\") pod \"oauth-openshift-5d4455474f-tnwpf\" (UID: \"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a\") " pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.757469 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a-audit-dir\") pod \"oauth-openshift-5d4455474f-tnwpf\" (UID: \"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a\") " pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.757685 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a-audit-policies\") pod \"oauth-openshift-5d4455474f-tnwpf\" (UID: \"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a\") " pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.757771 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5d4455474f-tnwpf\" (UID: \"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a\") " pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.758418 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5d4455474f-tnwpf\" (UID: \"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a\") " pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.763632 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a-v4-0-config-user-template-error\") pod \"oauth-openshift-5d4455474f-tnwpf\" (UID: \"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a\") " pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.763699 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5d4455474f-tnwpf\" (UID: \"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a\") " pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.763658 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a-v4-0-config-system-router-certs\") pod \"oauth-openshift-5d4455474f-tnwpf\" (UID: \"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a\") " pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.763653 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a-v4-0-config-user-template-login\") pod \"oauth-openshift-5d4455474f-tnwpf\" (UID: \"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a\") " pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.764302 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a-v4-0-config-system-session\") pod \"oauth-openshift-5d4455474f-tnwpf\" (UID: \"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a\") " pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.766338 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5d4455474f-tnwpf\" (UID: \"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a\") " pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.767292 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5d4455474f-tnwpf\" (UID: \"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a\") " pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.767301 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5d4455474f-tnwpf\" (UID: \"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a\") " pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.781141 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.787932 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-956kp\" (UniqueName: \"kubernetes.io/projected/34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a-kube-api-access-956kp\") pod \"oauth-openshift-5d4455474f-tnwpf\" (UID: \"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a\") " pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.829431 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.841829 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.890096 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:39 crc kubenswrapper[4856]: I0126 17:03:39.904548 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Jan 26 17:03:40 crc kubenswrapper[4856]: I0126 17:03:40.097032 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Jan 26 17:03:40 crc kubenswrapper[4856]: I0126 17:03:40.111477 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Jan 26 17:03:40 crc kubenswrapper[4856]: I0126 17:03:40.207150 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Jan 26 17:03:40 crc kubenswrapper[4856]: I0126 17:03:40.207510 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Jan 26 17:03:40 crc kubenswrapper[4856]: I0126 17:03:40.292644 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Jan 26 17:03:40 crc kubenswrapper[4856]: I0126 17:03:40.336621 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Jan 26 17:03:40 crc kubenswrapper[4856]: I0126 17:03:40.352045 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 26 17:03:40 crc kubenswrapper[4856]: I0126 17:03:40.379848 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Jan 26 17:03:40 crc kubenswrapper[4856]: I0126 17:03:40.386862 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Jan 26 17:03:40 crc kubenswrapper[4856]: I0126 17:03:40.437553 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Jan 26 17:03:40 crc kubenswrapper[4856]: I0126 17:03:40.446682 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Jan 26 17:03:40 crc kubenswrapper[4856]: I0126 17:03:40.534301 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Jan 26 17:03:40 crc kubenswrapper[4856]: I0126 17:03:40.590842 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Jan 26 17:03:40 crc kubenswrapper[4856]: I0126 17:03:40.614767 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Jan 26 17:03:40 crc kubenswrapper[4856]: I0126 17:03:40.653362 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Jan 26 17:03:40 crc kubenswrapper[4856]: I0126 17:03:40.673389 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 26 17:03:40 crc kubenswrapper[4856]: I0126 17:03:40.681718 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Jan 26 17:03:40 crc kubenswrapper[4856]: I0126 17:03:40.691009 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Jan 26 17:03:40 crc kubenswrapper[4856]: I0126 17:03:40.695944 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Jan 26 17:03:40 crc kubenswrapper[4856]: I0126 17:03:40.702046 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Jan 26 17:03:40 crc kubenswrapper[4856]: I0126 17:03:40.811949 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 26 17:03:40 crc kubenswrapper[4856]: I0126 17:03:40.920564 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Jan 26 17:03:40 crc kubenswrapper[4856]: I0126 17:03:40.993421 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Jan 26 17:03:41 crc kubenswrapper[4856]: I0126 17:03:41.070145 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Jan 26 17:03:41 crc kubenswrapper[4856]: I0126 17:03:41.099006 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Jan 26 17:03:41 crc kubenswrapper[4856]: I0126 17:03:41.114802 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Jan 26 17:03:41 crc kubenswrapper[4856]: I0126 17:03:41.207020 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Jan 26 17:03:41 crc kubenswrapper[4856]: I0126 17:03:41.235459 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Jan 26 17:03:41 crc kubenswrapper[4856]: I0126 17:03:41.386121 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 26 17:03:41 crc kubenswrapper[4856]: I0126 17:03:41.401655 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69008ed1-f3e5-400d-852f-adbcd94199f6" path="/var/lib/kubelet/pods/69008ed1-f3e5-400d-852f-adbcd94199f6/volumes" Jan 26 17:03:41 crc kubenswrapper[4856]: I0126 17:03:41.402569 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfa40861-cc08-4145-a185-6a3fb07eaabe" path="/var/lib/kubelet/pods/cfa40861-cc08-4145-a185-6a3fb07eaabe/volumes" Jan 26 17:03:41 crc kubenswrapper[4856]: I0126 17:03:41.507581 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Jan 26 17:03:41 crc kubenswrapper[4856]: I0126 17:03:41.552775 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Jan 26 17:03:41 crc kubenswrapper[4856]: I0126 17:03:41.676293 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Jan 26 17:03:41 crc kubenswrapper[4856]: I0126 17:03:41.681606 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Jan 26 17:03:41 crc kubenswrapper[4856]: I0126 17:03:41.768400 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Jan 26 17:03:41 crc kubenswrapper[4856]: I0126 17:03:41.774815 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Jan 26 17:03:41 crc kubenswrapper[4856]: I0126 17:03:41.784967 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Jan 26 17:03:41 crc kubenswrapper[4856]: I0126 17:03:41.856110 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Jan 26 17:03:41 crc kubenswrapper[4856]: I0126 17:03:41.869392 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 26 17:03:41 crc kubenswrapper[4856]: I0126 17:03:41.870980 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 26 17:03:41 crc kubenswrapper[4856]: I0126 17:03:41.967208 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Jan 26 17:03:42 crc kubenswrapper[4856]: I0126 17:03:42.090905 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Jan 26 17:03:42 crc kubenswrapper[4856]: I0126 17:03:42.100956 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Jan 26 17:03:42 crc kubenswrapper[4856]: I0126 17:03:42.117194 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Jan 26 17:03:42 crc kubenswrapper[4856]: I0126 17:03:42.201162 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Jan 26 17:03:42 crc kubenswrapper[4856]: I0126 17:03:42.427050 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Jan 26 17:03:42 crc kubenswrapper[4856]: I0126 17:03:42.444486 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Jan 26 17:03:42 crc kubenswrapper[4856]: I0126 17:03:42.455579 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Jan 26 17:03:42 crc kubenswrapper[4856]: I0126 17:03:42.495994 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Jan 26 17:03:42 crc kubenswrapper[4856]: I0126 17:03:42.511041 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Jan 26 17:03:42 crc kubenswrapper[4856]: I0126 17:03:42.572175 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Jan 26 17:03:42 crc kubenswrapper[4856]: I0126 17:03:42.600918 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Jan 26 17:03:42 crc kubenswrapper[4856]: I0126 17:03:42.731826 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Jan 26 17:03:42 crc kubenswrapper[4856]: I0126 17:03:42.739656 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Jan 26 17:03:42 crc kubenswrapper[4856]: I0126 17:03:42.744153 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Jan 26 17:03:42 crc kubenswrapper[4856]: I0126 17:03:42.750278 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 26 17:03:42 crc kubenswrapper[4856]: I0126 17:03:42.784743 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Jan 26 17:03:42 crc kubenswrapper[4856]: I0126 17:03:42.826290 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Jan 26 17:03:42 crc kubenswrapper[4856]: I0126 17:03:42.847334 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Jan 26 17:03:43 crc kubenswrapper[4856]: I0126 17:03:43.106890 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Jan 26 17:03:43 crc kubenswrapper[4856]: I0126 17:03:43.273916 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Jan 26 17:03:43 crc kubenswrapper[4856]: I0126 17:03:43.344054 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Jan 26 17:03:43 crc kubenswrapper[4856]: I0126 17:03:43.361784 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Jan 26 17:03:43 crc kubenswrapper[4856]: I0126 17:03:43.413162 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Jan 26 17:03:43 crc kubenswrapper[4856]: I0126 17:03:43.448067 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Jan 26 17:03:43 crc kubenswrapper[4856]: I0126 17:03:43.467016 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 26 17:03:43 crc kubenswrapper[4856]: I0126 17:03:43.572175 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Jan 26 17:03:43 crc kubenswrapper[4856]: I0126 17:03:43.615051 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Jan 26 17:03:43 crc kubenswrapper[4856]: I0126 17:03:43.691321 4856 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Jan 26 17:03:43 crc kubenswrapper[4856]: I0126 17:03:43.707851 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Jan 26 17:03:43 crc kubenswrapper[4856]: I0126 17:03:43.846910 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Jan 26 17:03:43 crc kubenswrapper[4856]: I0126 17:03:43.882330 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Jan 26 17:03:43 crc kubenswrapper[4856]: I0126 17:03:43.890218 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Jan 26 17:03:43 crc kubenswrapper[4856]: I0126 17:03:43.892283 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Jan 26 17:03:44 crc kubenswrapper[4856]: I0126 17:03:44.057943 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Jan 26 17:03:44 crc kubenswrapper[4856]: I0126 17:03:44.063123 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Jan 26 17:03:44 crc kubenswrapper[4856]: I0126 17:03:44.116621 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Jan 26 17:03:44 crc kubenswrapper[4856]: I0126 17:03:44.184034 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Jan 26 17:03:44 crc kubenswrapper[4856]: I0126 17:03:44.187346 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 26 17:03:44 crc kubenswrapper[4856]: I0126 17:03:44.240205 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Jan 26 17:03:44 crc kubenswrapper[4856]: I0126 17:03:44.296384 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Jan 26 17:03:44 crc kubenswrapper[4856]: I0126 17:03:44.302847 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Jan 26 17:03:44 crc kubenswrapper[4856]: I0126 17:03:44.368562 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 26 17:03:44 crc kubenswrapper[4856]: I0126 17:03:44.410473 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Jan 26 17:03:44 crc kubenswrapper[4856]: I0126 17:03:44.444545 4856 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 17:03:44 crc kubenswrapper[4856]: I0126 17:03:44.444594 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 17:03:44 crc kubenswrapper[4856]: I0126 17:03:44.450655 4856 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 17:03:44 crc kubenswrapper[4856]: I0126 17:03:44.458659 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Jan 26 17:03:44 crc kubenswrapper[4856]: I0126 17:03:44.509055 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Jan 26 17:03:44 crc kubenswrapper[4856]: I0126 17:03:44.619013 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Jan 26 17:03:44 crc kubenswrapper[4856]: I0126 17:03:44.628947 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 17:03:44 crc kubenswrapper[4856]: I0126 17:03:44.642602 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Jan 26 17:03:44 crc kubenswrapper[4856]: I0126 17:03:44.669216 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Jan 26 17:03:44 crc kubenswrapper[4856]: I0126 17:03:44.717699 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Jan 26 17:03:44 crc kubenswrapper[4856]: I0126 17:03:44.718203 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Jan 26 17:03:44 crc kubenswrapper[4856]: I0126 17:03:44.724828 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Jan 26 17:03:44 crc kubenswrapper[4856]: I0126 17:03:44.886059 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Jan 26 17:03:44 crc kubenswrapper[4856]: I0126 17:03:44.918692 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Jan 26 17:03:45 crc kubenswrapper[4856]: I0126 17:03:45.006189 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Jan 26 17:03:45 crc kubenswrapper[4856]: I0126 17:03:45.006211 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Jan 26 17:03:45 crc kubenswrapper[4856]: I0126 17:03:45.032771 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 26 17:03:45 crc kubenswrapper[4856]: I0126 17:03:45.043484 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Jan 26 17:03:45 crc kubenswrapper[4856]: I0126 17:03:45.155896 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Jan 26 17:03:45 crc kubenswrapper[4856]: I0126 17:03:45.192913 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Jan 26 17:03:45 crc kubenswrapper[4856]: I0126 17:03:45.220787 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Jan 26 17:03:45 crc kubenswrapper[4856]: I0126 17:03:45.252971 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Jan 26 17:03:45 crc kubenswrapper[4856]: I0126 17:03:45.349091 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Jan 26 17:03:45 crc kubenswrapper[4856]: I0126 17:03:45.638498 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Jan 26 17:03:45 crc kubenswrapper[4856]: I0126 17:03:45.748334 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Jan 26 17:03:45 crc kubenswrapper[4856]: I0126 17:03:45.819406 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Jan 26 17:03:45 crc kubenswrapper[4856]: I0126 17:03:45.858922 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Jan 26 17:03:45 crc kubenswrapper[4856]: I0126 17:03:45.913757 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Jan 26 17:03:45 crc kubenswrapper[4856]: I0126 17:03:45.933672 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Jan 26 17:03:45 crc kubenswrapper[4856]: I0126 17:03:45.957838 4856 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Jan 26 17:03:46 crc kubenswrapper[4856]: I0126 17:03:46.001944 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 26 17:03:46 crc kubenswrapper[4856]: I0126 17:03:46.075863 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Jan 26 17:03:46 crc kubenswrapper[4856]: I0126 17:03:46.222434 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Jan 26 17:03:46 crc kubenswrapper[4856]: I0126 17:03:46.398323 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Jan 26 17:03:46 crc kubenswrapper[4856]: I0126 17:03:46.401310 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5d4455474f-tnwpf"] Jan 26 17:03:46 crc kubenswrapper[4856]: I0126 17:03:46.436347 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Jan 26 17:03:46 crc kubenswrapper[4856]: I0126 17:03:46.487373 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Jan 26 17:03:46 crc kubenswrapper[4856]: I0126 17:03:46.570503 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5d4455474f-tnwpf"] Jan 26 17:03:46 crc kubenswrapper[4856]: I0126 17:03:46.629515 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Jan 26 17:03:46 crc kubenswrapper[4856]: I0126 17:03:46.636002 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Jan 26 17:03:46 crc kubenswrapper[4856]: I0126 17:03:46.729919 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Jan 26 17:03:46 crc kubenswrapper[4856]: I0126 17:03:46.730148 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Jan 26 17:03:46 crc kubenswrapper[4856]: I0126 17:03:46.730995 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Jan 26 17:03:46 crc kubenswrapper[4856]: I0126 17:03:46.741615 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" event={"ID":"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a","Type":"ContainerStarted","Data":"b0baa3f3e061df600bbe107f483216185a8feed4fc98462c2b563cfb8b2419a1"} Jan 26 17:03:46 crc kubenswrapper[4856]: I0126 17:03:46.784005 4856 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Jan 26 17:03:46 crc kubenswrapper[4856]: I0126 17:03:46.787840 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Jan 26 17:03:46 crc kubenswrapper[4856]: I0126 17:03:46.796830 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Jan 26 17:03:46 crc kubenswrapper[4856]: I0126 17:03:46.940772 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Jan 26 17:03:46 crc kubenswrapper[4856]: I0126 17:03:46.981682 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Jan 26 17:03:46 crc kubenswrapper[4856]: I0126 17:03:46.985889 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Jan 26 17:03:47 crc kubenswrapper[4856]: I0126 17:03:47.058127 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Jan 26 17:03:47 crc kubenswrapper[4856]: I0126 17:03:47.082881 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Jan 26 17:03:47 crc kubenswrapper[4856]: I0126 17:03:47.110882 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Jan 26 17:03:47 crc kubenswrapper[4856]: I0126 17:03:47.232009 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Jan 26 17:03:47 crc kubenswrapper[4856]: I0126 17:03:47.271316 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Jan 26 17:03:47 crc kubenswrapper[4856]: I0126 17:03:47.391287 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Jan 26 17:03:47 crc kubenswrapper[4856]: I0126 17:03:47.486696 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Jan 26 17:03:47 crc kubenswrapper[4856]: I0126 17:03:47.561234 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Jan 26 17:03:47 crc kubenswrapper[4856]: I0126 17:03:47.592368 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Jan 26 17:03:47 crc kubenswrapper[4856]: I0126 17:03:47.749472 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" event={"ID":"34c90e28-cdc1-4cd3-a7e1-d84a17c9ce5a","Type":"ContainerStarted","Data":"ec41a4e78361c1dbb8d9be4c452cd1e41744d581006301daa2c79b9405a0c8c1"} Jan 26 17:03:47 crc kubenswrapper[4856]: I0126 17:03:47.749945 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:47 crc kubenswrapper[4856]: I0126 17:03:47.755096 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" Jan 26 17:03:47 crc kubenswrapper[4856]: I0126 17:03:47.769381 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-5d4455474f-tnwpf" podStartSLOduration=63.769357411 podStartE2EDuration="1m3.769357411s" podCreationTimestamp="2026-01-26 17:02:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:03:47.768118315 +0000 UTC m=+323.721372316" watchObservedRunningTime="2026-01-26 17:03:47.769357411 +0000 UTC m=+323.722611392" Jan 26 17:03:47 crc kubenswrapper[4856]: I0126 17:03:47.850610 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Jan 26 17:03:47 crc kubenswrapper[4856]: I0126 17:03:47.908322 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Jan 26 17:03:47 crc kubenswrapper[4856]: I0126 17:03:47.996729 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Jan 26 17:03:48 crc kubenswrapper[4856]: I0126 17:03:48.060113 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 26 17:03:48 crc kubenswrapper[4856]: I0126 17:03:48.084512 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Jan 26 17:03:48 crc kubenswrapper[4856]: I0126 17:03:48.272758 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Jan 26 17:03:48 crc kubenswrapper[4856]: I0126 17:03:48.335245 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 26 17:03:48 crc kubenswrapper[4856]: I0126 17:03:48.595865 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Jan 26 17:03:55 crc kubenswrapper[4856]: I0126 17:03:55.541564 4856 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 26 17:03:55 crc kubenswrapper[4856]: I0126 17:03:55.541963 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://b6f864463a7443541f4490597f84a8f03e6d8c1a587e47002bb795632c0df2d6" gracePeriod=5 Jan 26 17:04:01 crc kubenswrapper[4856]: I0126 17:04:01.108361 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 26 17:04:01 crc kubenswrapper[4856]: I0126 17:04:01.108817 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 17:04:01 crc kubenswrapper[4856]: I0126 17:04:01.157088 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 26 17:04:01 crc kubenswrapper[4856]: I0126 17:04:01.157155 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 26 17:04:01 crc kubenswrapper[4856]: I0126 17:04:01.157182 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 26 17:04:01 crc kubenswrapper[4856]: I0126 17:04:01.157206 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 26 17:04:01 crc kubenswrapper[4856]: I0126 17:04:01.157238 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 26 17:04:01 crc kubenswrapper[4856]: I0126 17:04:01.157266 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:04:01 crc kubenswrapper[4856]: I0126 17:04:01.157302 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:04:01 crc kubenswrapper[4856]: I0126 17:04:01.157282 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:04:01 crc kubenswrapper[4856]: I0126 17:04:01.157241 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:04:01 crc kubenswrapper[4856]: I0126 17:04:01.157550 4856 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Jan 26 17:04:01 crc kubenswrapper[4856]: I0126 17:04:01.157568 4856 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 26 17:04:01 crc kubenswrapper[4856]: I0126 17:04:01.157580 4856 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Jan 26 17:04:01 crc kubenswrapper[4856]: I0126 17:04:01.157593 4856 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Jan 26 17:04:01 crc kubenswrapper[4856]: I0126 17:04:01.165293 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:04:01 crc kubenswrapper[4856]: I0126 17:04:01.258414 4856 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 26 17:04:01 crc kubenswrapper[4856]: I0126 17:04:01.406497 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Jan 26 17:04:01 crc kubenswrapper[4856]: I0126 17:04:01.407078 4856 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Jan 26 17:04:01 crc kubenswrapper[4856]: I0126 17:04:01.436870 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 26 17:04:01 crc kubenswrapper[4856]: I0126 17:04:01.436929 4856 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="b2d37eac-1f35-4184-aaf4-fe3e28069de2" Jan 26 17:04:01 crc kubenswrapper[4856]: I0126 17:04:01.439550 4856 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 26 17:04:01 crc kubenswrapper[4856]: I0126 17:04:01.439598 4856 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="b2d37eac-1f35-4184-aaf4-fe3e28069de2" Jan 26 17:04:01 crc kubenswrapper[4856]: I0126 17:04:01.840018 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 26 17:04:01 crc kubenswrapper[4856]: I0126 17:04:01.840326 4856 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="b6f864463a7443541f4490597f84a8f03e6d8c1a587e47002bb795632c0df2d6" exitCode=137 Jan 26 17:04:01 crc kubenswrapper[4856]: I0126 17:04:01.840389 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 17:04:01 crc kubenswrapper[4856]: I0126 17:04:01.840426 4856 scope.go:117] "RemoveContainer" containerID="b6f864463a7443541f4490597f84a8f03e6d8c1a587e47002bb795632c0df2d6" Jan 26 17:04:01 crc kubenswrapper[4856]: I0126 17:04:01.859307 4856 scope.go:117] "RemoveContainer" containerID="b6f864463a7443541f4490597f84a8f03e6d8c1a587e47002bb795632c0df2d6" Jan 26 17:04:01 crc kubenswrapper[4856]: E0126 17:04:01.859880 4856 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6f864463a7443541f4490597f84a8f03e6d8c1a587e47002bb795632c0df2d6\": container with ID starting with b6f864463a7443541f4490597f84a8f03e6d8c1a587e47002bb795632c0df2d6 not found: ID does not exist" containerID="b6f864463a7443541f4490597f84a8f03e6d8c1a587e47002bb795632c0df2d6" Jan 26 17:04:01 crc kubenswrapper[4856]: I0126 17:04:01.859944 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6f864463a7443541f4490597f84a8f03e6d8c1a587e47002bb795632c0df2d6"} err="failed to get container status \"b6f864463a7443541f4490597f84a8f03e6d8c1a587e47002bb795632c0df2d6\": rpc error: code = NotFound desc = could not find container \"b6f864463a7443541f4490597f84a8f03e6d8c1a587e47002bb795632c0df2d6\": container with ID starting with b6f864463a7443541f4490597f84a8f03e6d8c1a587e47002bb795632c0df2d6 not found: ID does not exist" Jan 26 17:04:09 crc kubenswrapper[4856]: I0126 17:04:09.477776 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-54b988dd69-ljwqg"] Jan 26 17:04:09 crc kubenswrapper[4856]: I0126 17:04:09.478636 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-54b988dd69-ljwqg" podUID="de17aec3-fab1-4a5e-bd46-6a1545b93a89" containerName="controller-manager" containerID="cri-o://5dd5f652c6d735efc7f4d83a862afc4f09c88afb0921e6ecb304af76698cc9c8" gracePeriod=30 Jan 26 17:04:09 crc kubenswrapper[4856]: I0126 17:04:09.578665 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cb449784d-bqprm"] Jan 26 17:04:09 crc kubenswrapper[4856]: I0126 17:04:09.578908 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-cb449784d-bqprm" podUID="54dee8cd-259a-4c9f-9e56-fbd0ea167f46" containerName="route-controller-manager" containerID="cri-o://3157d6dd787fe30eefd4db24c3f3619f52444746080c65c699eab5d0d02ab52a" gracePeriod=30 Jan 26 17:04:09 crc kubenswrapper[4856]: I0126 17:04:09.911329 4856 generic.go:334] "Generic (PLEG): container finished" podID="de17aec3-fab1-4a5e-bd46-6a1545b93a89" containerID="5dd5f652c6d735efc7f4d83a862afc4f09c88afb0921e6ecb304af76698cc9c8" exitCode=0 Jan 26 17:04:09 crc kubenswrapper[4856]: I0126 17:04:09.911805 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-54b988dd69-ljwqg" event={"ID":"de17aec3-fab1-4a5e-bd46-6a1545b93a89","Type":"ContainerDied","Data":"5dd5f652c6d735efc7f4d83a862afc4f09c88afb0921e6ecb304af76698cc9c8"} Jan 26 17:04:09 crc kubenswrapper[4856]: I0126 17:04:09.914534 4856 generic.go:334] "Generic (PLEG): container finished" podID="54dee8cd-259a-4c9f-9e56-fbd0ea167f46" containerID="3157d6dd787fe30eefd4db24c3f3619f52444746080c65c699eab5d0d02ab52a" exitCode=0 Jan 26 17:04:09 crc kubenswrapper[4856]: I0126 17:04:09.914615 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-cb449784d-bqprm" event={"ID":"54dee8cd-259a-4c9f-9e56-fbd0ea167f46","Type":"ContainerDied","Data":"3157d6dd787fe30eefd4db24c3f3619f52444746080c65c699eab5d0d02ab52a"} Jan 26 17:04:09 crc kubenswrapper[4856]: I0126 17:04:09.948624 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-54b988dd69-ljwqg" Jan 26 17:04:10 crc kubenswrapper[4856]: I0126 17:04:10.070746 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-cb449784d-bqprm" Jan 26 17:04:10 crc kubenswrapper[4856]: I0126 17:04:10.106763 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sxzpz\" (UniqueName: \"kubernetes.io/projected/de17aec3-fab1-4a5e-bd46-6a1545b93a89-kube-api-access-sxzpz\") pod \"de17aec3-fab1-4a5e-bd46-6a1545b93a89\" (UID: \"de17aec3-fab1-4a5e-bd46-6a1545b93a89\") " Jan 26 17:04:10 crc kubenswrapper[4856]: I0126 17:04:10.106849 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/de17aec3-fab1-4a5e-bd46-6a1545b93a89-proxy-ca-bundles\") pod \"de17aec3-fab1-4a5e-bd46-6a1545b93a89\" (UID: \"de17aec3-fab1-4a5e-bd46-6a1545b93a89\") " Jan 26 17:04:10 crc kubenswrapper[4856]: I0126 17:04:10.106893 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/de17aec3-fab1-4a5e-bd46-6a1545b93a89-serving-cert\") pod \"de17aec3-fab1-4a5e-bd46-6a1545b93a89\" (UID: \"de17aec3-fab1-4a5e-bd46-6a1545b93a89\") " Jan 26 17:04:10 crc kubenswrapper[4856]: I0126 17:04:10.106926 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/de17aec3-fab1-4a5e-bd46-6a1545b93a89-client-ca\") pod \"de17aec3-fab1-4a5e-bd46-6a1545b93a89\" (UID: \"de17aec3-fab1-4a5e-bd46-6a1545b93a89\") " Jan 26 17:04:10 crc kubenswrapper[4856]: I0126 17:04:10.107680 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de17aec3-fab1-4a5e-bd46-6a1545b93a89-client-ca" (OuterVolumeSpecName: "client-ca") pod "de17aec3-fab1-4a5e-bd46-6a1545b93a89" (UID: "de17aec3-fab1-4a5e-bd46-6a1545b93a89"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:04:10 crc kubenswrapper[4856]: I0126 17:04:10.107706 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de17aec3-fab1-4a5e-bd46-6a1545b93a89-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "de17aec3-fab1-4a5e-bd46-6a1545b93a89" (UID: "de17aec3-fab1-4a5e-bd46-6a1545b93a89"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:04:10 crc kubenswrapper[4856]: I0126 17:04:10.106982 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de17aec3-fab1-4a5e-bd46-6a1545b93a89-config\") pod \"de17aec3-fab1-4a5e-bd46-6a1545b93a89\" (UID: \"de17aec3-fab1-4a5e-bd46-6a1545b93a89\") " Jan 26 17:04:10 crc kubenswrapper[4856]: I0126 17:04:10.107908 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de17aec3-fab1-4a5e-bd46-6a1545b93a89-config" (OuterVolumeSpecName: "config") pod "de17aec3-fab1-4a5e-bd46-6a1545b93a89" (UID: "de17aec3-fab1-4a5e-bd46-6a1545b93a89"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:04:10 crc kubenswrapper[4856]: I0126 17:04:10.108501 4856 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/de17aec3-fab1-4a5e-bd46-6a1545b93a89-client-ca\") on node \"crc\" DevicePath \"\"" Jan 26 17:04:10 crc kubenswrapper[4856]: I0126 17:04:10.108542 4856 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de17aec3-fab1-4a5e-bd46-6a1545b93a89-config\") on node \"crc\" DevicePath \"\"" Jan 26 17:04:10 crc kubenswrapper[4856]: I0126 17:04:10.108561 4856 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/de17aec3-fab1-4a5e-bd46-6a1545b93a89-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 26 17:04:10 crc kubenswrapper[4856]: I0126 17:04:10.111941 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de17aec3-fab1-4a5e-bd46-6a1545b93a89-kube-api-access-sxzpz" (OuterVolumeSpecName: "kube-api-access-sxzpz") pod "de17aec3-fab1-4a5e-bd46-6a1545b93a89" (UID: "de17aec3-fab1-4a5e-bd46-6a1545b93a89"). InnerVolumeSpecName "kube-api-access-sxzpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:04:10 crc kubenswrapper[4856]: I0126 17:04:10.112707 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de17aec3-fab1-4a5e-bd46-6a1545b93a89-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "de17aec3-fab1-4a5e-bd46-6a1545b93a89" (UID: "de17aec3-fab1-4a5e-bd46-6a1545b93a89"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:04:10 crc kubenswrapper[4856]: I0126 17:04:10.209490 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54dee8cd-259a-4c9f-9e56-fbd0ea167f46-config\") pod \"54dee8cd-259a-4c9f-9e56-fbd0ea167f46\" (UID: \"54dee8cd-259a-4c9f-9e56-fbd0ea167f46\") " Jan 26 17:04:10 crc kubenswrapper[4856]: I0126 17:04:10.209566 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9vhl4\" (UniqueName: \"kubernetes.io/projected/54dee8cd-259a-4c9f-9e56-fbd0ea167f46-kube-api-access-9vhl4\") pod \"54dee8cd-259a-4c9f-9e56-fbd0ea167f46\" (UID: \"54dee8cd-259a-4c9f-9e56-fbd0ea167f46\") " Jan 26 17:04:10 crc kubenswrapper[4856]: I0126 17:04:10.209613 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54dee8cd-259a-4c9f-9e56-fbd0ea167f46-serving-cert\") pod \"54dee8cd-259a-4c9f-9e56-fbd0ea167f46\" (UID: \"54dee8cd-259a-4c9f-9e56-fbd0ea167f46\") " Jan 26 17:04:10 crc kubenswrapper[4856]: I0126 17:04:10.209656 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/54dee8cd-259a-4c9f-9e56-fbd0ea167f46-client-ca\") pod \"54dee8cd-259a-4c9f-9e56-fbd0ea167f46\" (UID: \"54dee8cd-259a-4c9f-9e56-fbd0ea167f46\") " Jan 26 17:04:10 crc kubenswrapper[4856]: I0126 17:04:10.209877 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sxzpz\" (UniqueName: \"kubernetes.io/projected/de17aec3-fab1-4a5e-bd46-6a1545b93a89-kube-api-access-sxzpz\") on node \"crc\" DevicePath \"\"" Jan 26 17:04:10 crc kubenswrapper[4856]: I0126 17:04:10.209894 4856 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/de17aec3-fab1-4a5e-bd46-6a1545b93a89-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 17:04:10 crc kubenswrapper[4856]: I0126 17:04:10.210695 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54dee8cd-259a-4c9f-9e56-fbd0ea167f46-config" (OuterVolumeSpecName: "config") pod "54dee8cd-259a-4c9f-9e56-fbd0ea167f46" (UID: "54dee8cd-259a-4c9f-9e56-fbd0ea167f46"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:04:10 crc kubenswrapper[4856]: I0126 17:04:10.210696 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54dee8cd-259a-4c9f-9e56-fbd0ea167f46-client-ca" (OuterVolumeSpecName: "client-ca") pod "54dee8cd-259a-4c9f-9e56-fbd0ea167f46" (UID: "54dee8cd-259a-4c9f-9e56-fbd0ea167f46"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:04:10 crc kubenswrapper[4856]: I0126 17:04:10.212753 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54dee8cd-259a-4c9f-9e56-fbd0ea167f46-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "54dee8cd-259a-4c9f-9e56-fbd0ea167f46" (UID: "54dee8cd-259a-4c9f-9e56-fbd0ea167f46"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:04:10 crc kubenswrapper[4856]: I0126 17:04:10.215141 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54dee8cd-259a-4c9f-9e56-fbd0ea167f46-kube-api-access-9vhl4" (OuterVolumeSpecName: "kube-api-access-9vhl4") pod "54dee8cd-259a-4c9f-9e56-fbd0ea167f46" (UID: "54dee8cd-259a-4c9f-9e56-fbd0ea167f46"). InnerVolumeSpecName "kube-api-access-9vhl4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:04:10 crc kubenswrapper[4856]: I0126 17:04:10.310709 4856 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54dee8cd-259a-4c9f-9e56-fbd0ea167f46-config\") on node \"crc\" DevicePath \"\"" Jan 26 17:04:10 crc kubenswrapper[4856]: I0126 17:04:10.310751 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9vhl4\" (UniqueName: \"kubernetes.io/projected/54dee8cd-259a-4c9f-9e56-fbd0ea167f46-kube-api-access-9vhl4\") on node \"crc\" DevicePath \"\"" Jan 26 17:04:10 crc kubenswrapper[4856]: I0126 17:04:10.310763 4856 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54dee8cd-259a-4c9f-9e56-fbd0ea167f46-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 17:04:10 crc kubenswrapper[4856]: I0126 17:04:10.310777 4856 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/54dee8cd-259a-4c9f-9e56-fbd0ea167f46-client-ca\") on node \"crc\" DevicePath \"\"" Jan 26 17:04:10 crc kubenswrapper[4856]: I0126 17:04:10.921181 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-cb449784d-bqprm" event={"ID":"54dee8cd-259a-4c9f-9e56-fbd0ea167f46","Type":"ContainerDied","Data":"960ff00bedf28636eb04c4f352e2d6d2e33a5ceb9800e901e018103cd5ac5859"} Jan 26 17:04:10 crc kubenswrapper[4856]: I0126 17:04:10.921213 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-cb449784d-bqprm" Jan 26 17:04:10 crc kubenswrapper[4856]: I0126 17:04:10.921256 4856 scope.go:117] "RemoveContainer" containerID="3157d6dd787fe30eefd4db24c3f3619f52444746080c65c699eab5d0d02ab52a" Jan 26 17:04:10 crc kubenswrapper[4856]: I0126 17:04:10.924919 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-54b988dd69-ljwqg" event={"ID":"de17aec3-fab1-4a5e-bd46-6a1545b93a89","Type":"ContainerDied","Data":"1bccd3720f328e7d0b92fc36bcc35726a97ebe6a8070f5cbb1608de57071e2d0"} Jan 26 17:04:10 crc kubenswrapper[4856]: I0126 17:04:10.925083 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-54b988dd69-ljwqg" Jan 26 17:04:10 crc kubenswrapper[4856]: I0126 17:04:10.940678 4856 scope.go:117] "RemoveContainer" containerID="5dd5f652c6d735efc7f4d83a862afc4f09c88afb0921e6ecb304af76698cc9c8" Jan 26 17:04:10 crc kubenswrapper[4856]: I0126 17:04:10.950049 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cb449784d-bqprm"] Jan 26 17:04:10 crc kubenswrapper[4856]: I0126 17:04:10.954906 4856 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cb449784d-bqprm"] Jan 26 17:04:10 crc kubenswrapper[4856]: I0126 17:04:10.958878 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-54b988dd69-ljwqg"] Jan 26 17:04:10 crc kubenswrapper[4856]: I0126 17:04:10.962133 4856 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-54b988dd69-ljwqg"] Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.394180 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-787548c685-7dskp"] Jan 26 17:04:11 crc kubenswrapper[4856]: E0126 17:04:11.394697 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54dee8cd-259a-4c9f-9e56-fbd0ea167f46" containerName="route-controller-manager" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.394720 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="54dee8cd-259a-4c9f-9e56-fbd0ea167f46" containerName="route-controller-manager" Jan 26 17:04:11 crc kubenswrapper[4856]: E0126 17:04:11.394739 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de17aec3-fab1-4a5e-bd46-6a1545b93a89" containerName="controller-manager" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.394749 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="de17aec3-fab1-4a5e-bd46-6a1545b93a89" containerName="controller-manager" Jan 26 17:04:11 crc kubenswrapper[4856]: E0126 17:04:11.394775 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.394786 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.395050 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.395071 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="de17aec3-fab1-4a5e-bd46-6a1545b93a89" containerName="controller-manager" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.395095 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="54dee8cd-259a-4c9f-9e56-fbd0ea167f46" containerName="route-controller-manager" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.399641 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-787548c685-7dskp" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.403559 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.403809 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.403891 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.404217 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.404413 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.404474 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.405148 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54dee8cd-259a-4c9f-9e56-fbd0ea167f46" path="/var/lib/kubelet/pods/54dee8cd-259a-4c9f-9e56-fbd0ea167f46/volumes" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.406170 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de17aec3-fab1-4a5e-bd46-6a1545b93a89" path="/var/lib/kubelet/pods/de17aec3-fab1-4a5e-bd46-6a1545b93a89/volumes" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.406857 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c4b475647-2td8w"] Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.407884 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c4b475647-2td8w"] Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.408021 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5c4b475647-2td8w" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.411077 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.411276 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.411440 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.411497 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.411756 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.420907 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.421852 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-787548c685-7dskp"] Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.423782 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.527105 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f710888e-8c73-4d02-8ab4-f530b2562d8e-proxy-ca-bundles\") pod \"controller-manager-787548c685-7dskp\" (UID: \"f710888e-8c73-4d02-8ab4-f530b2562d8e\") " pod="openshift-controller-manager/controller-manager-787548c685-7dskp" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.527157 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f710888e-8c73-4d02-8ab4-f530b2562d8e-client-ca\") pod \"controller-manager-787548c685-7dskp\" (UID: \"f710888e-8c73-4d02-8ab4-f530b2562d8e\") " pod="openshift-controller-manager/controller-manager-787548c685-7dskp" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.527576 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4edd939-e622-4a82-9c29-a115042d697e-config\") pod \"route-controller-manager-5c4b475647-2td8w\" (UID: \"e4edd939-e622-4a82-9c29-a115042d697e\") " pod="openshift-route-controller-manager/route-controller-manager-5c4b475647-2td8w" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.527656 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7x9xc\" (UniqueName: \"kubernetes.io/projected/f710888e-8c73-4d02-8ab4-f530b2562d8e-kube-api-access-7x9xc\") pod \"controller-manager-787548c685-7dskp\" (UID: \"f710888e-8c73-4d02-8ab4-f530b2562d8e\") " pod="openshift-controller-manager/controller-manager-787548c685-7dskp" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.527773 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f710888e-8c73-4d02-8ab4-f530b2562d8e-config\") pod \"controller-manager-787548c685-7dskp\" (UID: \"f710888e-8c73-4d02-8ab4-f530b2562d8e\") " pod="openshift-controller-manager/controller-manager-787548c685-7dskp" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.527830 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnhw8\" (UniqueName: \"kubernetes.io/projected/e4edd939-e622-4a82-9c29-a115042d697e-kube-api-access-dnhw8\") pod \"route-controller-manager-5c4b475647-2td8w\" (UID: \"e4edd939-e622-4a82-9c29-a115042d697e\") " pod="openshift-route-controller-manager/route-controller-manager-5c4b475647-2td8w" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.527860 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e4edd939-e622-4a82-9c29-a115042d697e-serving-cert\") pod \"route-controller-manager-5c4b475647-2td8w\" (UID: \"e4edd939-e622-4a82-9c29-a115042d697e\") " pod="openshift-route-controller-manager/route-controller-manager-5c4b475647-2td8w" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.527898 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e4edd939-e622-4a82-9c29-a115042d697e-client-ca\") pod \"route-controller-manager-5c4b475647-2td8w\" (UID: \"e4edd939-e622-4a82-9c29-a115042d697e\") " pod="openshift-route-controller-manager/route-controller-manager-5c4b475647-2td8w" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.527950 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f710888e-8c73-4d02-8ab4-f530b2562d8e-serving-cert\") pod \"controller-manager-787548c685-7dskp\" (UID: \"f710888e-8c73-4d02-8ab4-f530b2562d8e\") " pod="openshift-controller-manager/controller-manager-787548c685-7dskp" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.628841 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4edd939-e622-4a82-9c29-a115042d697e-config\") pod \"route-controller-manager-5c4b475647-2td8w\" (UID: \"e4edd939-e622-4a82-9c29-a115042d697e\") " pod="openshift-route-controller-manager/route-controller-manager-5c4b475647-2td8w" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.628898 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7x9xc\" (UniqueName: \"kubernetes.io/projected/f710888e-8c73-4d02-8ab4-f530b2562d8e-kube-api-access-7x9xc\") pod \"controller-manager-787548c685-7dskp\" (UID: \"f710888e-8c73-4d02-8ab4-f530b2562d8e\") " pod="openshift-controller-manager/controller-manager-787548c685-7dskp" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.628934 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f710888e-8c73-4d02-8ab4-f530b2562d8e-config\") pod \"controller-manager-787548c685-7dskp\" (UID: \"f710888e-8c73-4d02-8ab4-f530b2562d8e\") " pod="openshift-controller-manager/controller-manager-787548c685-7dskp" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.628988 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnhw8\" (UniqueName: \"kubernetes.io/projected/e4edd939-e622-4a82-9c29-a115042d697e-kube-api-access-dnhw8\") pod \"route-controller-manager-5c4b475647-2td8w\" (UID: \"e4edd939-e622-4a82-9c29-a115042d697e\") " pod="openshift-route-controller-manager/route-controller-manager-5c4b475647-2td8w" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.629012 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e4edd939-e622-4a82-9c29-a115042d697e-serving-cert\") pod \"route-controller-manager-5c4b475647-2td8w\" (UID: \"e4edd939-e622-4a82-9c29-a115042d697e\") " pod="openshift-route-controller-manager/route-controller-manager-5c4b475647-2td8w" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.629034 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e4edd939-e622-4a82-9c29-a115042d697e-client-ca\") pod \"route-controller-manager-5c4b475647-2td8w\" (UID: \"e4edd939-e622-4a82-9c29-a115042d697e\") " pod="openshift-route-controller-manager/route-controller-manager-5c4b475647-2td8w" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.629075 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f710888e-8c73-4d02-8ab4-f530b2562d8e-serving-cert\") pod \"controller-manager-787548c685-7dskp\" (UID: \"f710888e-8c73-4d02-8ab4-f530b2562d8e\") " pod="openshift-controller-manager/controller-manager-787548c685-7dskp" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.629102 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f710888e-8c73-4d02-8ab4-f530b2562d8e-proxy-ca-bundles\") pod \"controller-manager-787548c685-7dskp\" (UID: \"f710888e-8c73-4d02-8ab4-f530b2562d8e\") " pod="openshift-controller-manager/controller-manager-787548c685-7dskp" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.629128 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f710888e-8c73-4d02-8ab4-f530b2562d8e-client-ca\") pod \"controller-manager-787548c685-7dskp\" (UID: \"f710888e-8c73-4d02-8ab4-f530b2562d8e\") " pod="openshift-controller-manager/controller-manager-787548c685-7dskp" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.630168 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f710888e-8c73-4d02-8ab4-f530b2562d8e-client-ca\") pod \"controller-manager-787548c685-7dskp\" (UID: \"f710888e-8c73-4d02-8ab4-f530b2562d8e\") " pod="openshift-controller-manager/controller-manager-787548c685-7dskp" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.630242 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e4edd939-e622-4a82-9c29-a115042d697e-client-ca\") pod \"route-controller-manager-5c4b475647-2td8w\" (UID: \"e4edd939-e622-4a82-9c29-a115042d697e\") " pod="openshift-route-controller-manager/route-controller-manager-5c4b475647-2td8w" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.630398 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f710888e-8c73-4d02-8ab4-f530b2562d8e-config\") pod \"controller-manager-787548c685-7dskp\" (UID: \"f710888e-8c73-4d02-8ab4-f530b2562d8e\") " pod="openshift-controller-manager/controller-manager-787548c685-7dskp" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.630511 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4edd939-e622-4a82-9c29-a115042d697e-config\") pod \"route-controller-manager-5c4b475647-2td8w\" (UID: \"e4edd939-e622-4a82-9c29-a115042d697e\") " pod="openshift-route-controller-manager/route-controller-manager-5c4b475647-2td8w" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.633902 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f710888e-8c73-4d02-8ab4-f530b2562d8e-proxy-ca-bundles\") pod \"controller-manager-787548c685-7dskp\" (UID: \"f710888e-8c73-4d02-8ab4-f530b2562d8e\") " pod="openshift-controller-manager/controller-manager-787548c685-7dskp" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.634116 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f710888e-8c73-4d02-8ab4-f530b2562d8e-serving-cert\") pod \"controller-manager-787548c685-7dskp\" (UID: \"f710888e-8c73-4d02-8ab4-f530b2562d8e\") " pod="openshift-controller-manager/controller-manager-787548c685-7dskp" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.638206 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e4edd939-e622-4a82-9c29-a115042d697e-serving-cert\") pod \"route-controller-manager-5c4b475647-2td8w\" (UID: \"e4edd939-e622-4a82-9c29-a115042d697e\") " pod="openshift-route-controller-manager/route-controller-manager-5c4b475647-2td8w" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.644876 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7x9xc\" (UniqueName: \"kubernetes.io/projected/f710888e-8c73-4d02-8ab4-f530b2562d8e-kube-api-access-7x9xc\") pod \"controller-manager-787548c685-7dskp\" (UID: \"f710888e-8c73-4d02-8ab4-f530b2562d8e\") " pod="openshift-controller-manager/controller-manager-787548c685-7dskp" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.645847 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnhw8\" (UniqueName: \"kubernetes.io/projected/e4edd939-e622-4a82-9c29-a115042d697e-kube-api-access-dnhw8\") pod \"route-controller-manager-5c4b475647-2td8w\" (UID: \"e4edd939-e622-4a82-9c29-a115042d697e\") " pod="openshift-route-controller-manager/route-controller-manager-5c4b475647-2td8w" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.732945 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-787548c685-7dskp" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.745064 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5c4b475647-2td8w" Jan 26 17:04:11 crc kubenswrapper[4856]: I0126 17:04:11.960735 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c4b475647-2td8w"] Jan 26 17:04:12 crc kubenswrapper[4856]: I0126 17:04:12.004068 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-787548c685-7dskp"] Jan 26 17:04:12 crc kubenswrapper[4856]: W0126 17:04:12.008993 4856 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf710888e_8c73_4d02_8ab4_f530b2562d8e.slice/crio-558fe0d45eb56f6266e1083119bf551e438b00b093d926a293bf75ae0823f5a2 WatchSource:0}: Error finding container 558fe0d45eb56f6266e1083119bf551e438b00b093d926a293bf75ae0823f5a2: Status 404 returned error can't find the container with id 558fe0d45eb56f6266e1083119bf551e438b00b093d926a293bf75ae0823f5a2 Jan 26 17:04:12 crc kubenswrapper[4856]: I0126 17:04:12.941472 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5c4b475647-2td8w" event={"ID":"e4edd939-e622-4a82-9c29-a115042d697e","Type":"ContainerStarted","Data":"3686bdddd3134b3ea5d523f1b197d25f5fe8b842e6cb39fce2a80b12a3c66dbe"} Jan 26 17:04:12 crc kubenswrapper[4856]: I0126 17:04:12.941893 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5c4b475647-2td8w" event={"ID":"e4edd939-e622-4a82-9c29-a115042d697e","Type":"ContainerStarted","Data":"af1dee1a1d954458157b743dec4d52a5ba26c12dbe51718f2f422dd7e022e190"} Jan 26 17:04:12 crc kubenswrapper[4856]: I0126 17:04:12.941917 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5c4b475647-2td8w" Jan 26 17:04:12 crc kubenswrapper[4856]: I0126 17:04:12.943517 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-787548c685-7dskp" event={"ID":"f710888e-8c73-4d02-8ab4-f530b2562d8e","Type":"ContainerStarted","Data":"d59eb1b03a2fffa887687d53c3e17cf812845ab76f4c776e709172dc2d904988"} Jan 26 17:04:12 crc kubenswrapper[4856]: I0126 17:04:12.943564 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-787548c685-7dskp" event={"ID":"f710888e-8c73-4d02-8ab4-f530b2562d8e","Type":"ContainerStarted","Data":"558fe0d45eb56f6266e1083119bf551e438b00b093d926a293bf75ae0823f5a2"} Jan 26 17:04:12 crc kubenswrapper[4856]: I0126 17:04:12.943927 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-787548c685-7dskp" Jan 26 17:04:12 crc kubenswrapper[4856]: I0126 17:04:12.946825 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5c4b475647-2td8w" Jan 26 17:04:12 crc kubenswrapper[4856]: I0126 17:04:12.953973 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-787548c685-7dskp" Jan 26 17:04:12 crc kubenswrapper[4856]: I0126 17:04:12.960921 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5c4b475647-2td8w" podStartSLOduration=3.960893315 podStartE2EDuration="3.960893315s" podCreationTimestamp="2026-01-26 17:04:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:04:12.95869668 +0000 UTC m=+348.911950681" watchObservedRunningTime="2026-01-26 17:04:12.960893315 +0000 UTC m=+348.914147306" Jan 26 17:04:13 crc kubenswrapper[4856]: I0126 17:04:13.017621 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-787548c685-7dskp" podStartSLOduration=4.017588714 podStartE2EDuration="4.017588714s" podCreationTimestamp="2026-01-26 17:04:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:04:12.996200427 +0000 UTC m=+348.949454418" watchObservedRunningTime="2026-01-26 17:04:13.017588714 +0000 UTC m=+348.970842695" Jan 26 17:04:17 crc kubenswrapper[4856]: I0126 17:04:17.346750 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Jan 26 17:04:29 crc kubenswrapper[4856]: I0126 17:04:29.659973 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-787548c685-7dskp"] Jan 26 17:04:29 crc kubenswrapper[4856]: I0126 17:04:29.662332 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-787548c685-7dskp" podUID="f710888e-8c73-4d02-8ab4-f530b2562d8e" containerName="controller-manager" containerID="cri-o://d59eb1b03a2fffa887687d53c3e17cf812845ab76f4c776e709172dc2d904988" gracePeriod=30 Jan 26 17:04:30 crc kubenswrapper[4856]: I0126 17:04:30.052393 4856 generic.go:334] "Generic (PLEG): container finished" podID="f710888e-8c73-4d02-8ab4-f530b2562d8e" containerID="d59eb1b03a2fffa887687d53c3e17cf812845ab76f4c776e709172dc2d904988" exitCode=0 Jan 26 17:04:30 crc kubenswrapper[4856]: I0126 17:04:30.052456 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-787548c685-7dskp" event={"ID":"f710888e-8c73-4d02-8ab4-f530b2562d8e","Type":"ContainerDied","Data":"d59eb1b03a2fffa887687d53c3e17cf812845ab76f4c776e709172dc2d904988"} Jan 26 17:04:30 crc kubenswrapper[4856]: I0126 17:04:30.986122 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-787548c685-7dskp" Jan 26 17:04:31 crc kubenswrapper[4856]: I0126 17:04:31.019583 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5674dc9874-8hwcb"] Jan 26 17:04:31 crc kubenswrapper[4856]: E0126 17:04:31.020080 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f710888e-8c73-4d02-8ab4-f530b2562d8e" containerName="controller-manager" Jan 26 17:04:31 crc kubenswrapper[4856]: I0126 17:04:31.020107 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="f710888e-8c73-4d02-8ab4-f530b2562d8e" containerName="controller-manager" Jan 26 17:04:31 crc kubenswrapper[4856]: I0126 17:04:31.020285 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="f710888e-8c73-4d02-8ab4-f530b2562d8e" containerName="controller-manager" Jan 26 17:04:31 crc kubenswrapper[4856]: I0126 17:04:31.021012 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5674dc9874-8hwcb" Jan 26 17:04:31 crc kubenswrapper[4856]: I0126 17:04:31.031809 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5674dc9874-8hwcb"] Jan 26 17:04:31 crc kubenswrapper[4856]: I0126 17:04:31.063685 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-787548c685-7dskp" event={"ID":"f710888e-8c73-4d02-8ab4-f530b2562d8e","Type":"ContainerDied","Data":"558fe0d45eb56f6266e1083119bf551e438b00b093d926a293bf75ae0823f5a2"} Jan 26 17:04:31 crc kubenswrapper[4856]: I0126 17:04:31.063773 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-787548c685-7dskp" Jan 26 17:04:31 crc kubenswrapper[4856]: I0126 17:04:31.063827 4856 scope.go:117] "RemoveContainer" containerID="d59eb1b03a2fffa887687d53c3e17cf812845ab76f4c776e709172dc2d904988" Jan 26 17:04:31 crc kubenswrapper[4856]: I0126 17:04:31.121988 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f710888e-8c73-4d02-8ab4-f530b2562d8e-serving-cert\") pod \"f710888e-8c73-4d02-8ab4-f530b2562d8e\" (UID: \"f710888e-8c73-4d02-8ab4-f530b2562d8e\") " Jan 26 17:04:31 crc kubenswrapper[4856]: I0126 17:04:31.122147 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f710888e-8c73-4d02-8ab4-f530b2562d8e-config\") pod \"f710888e-8c73-4d02-8ab4-f530b2562d8e\" (UID: \"f710888e-8c73-4d02-8ab4-f530b2562d8e\") " Jan 26 17:04:31 crc kubenswrapper[4856]: I0126 17:04:31.122724 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f710888e-8c73-4d02-8ab4-f530b2562d8e-proxy-ca-bundles\") pod \"f710888e-8c73-4d02-8ab4-f530b2562d8e\" (UID: \"f710888e-8c73-4d02-8ab4-f530b2562d8e\") " Jan 26 17:04:31 crc kubenswrapper[4856]: I0126 17:04:31.122760 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f710888e-8c73-4d02-8ab4-f530b2562d8e-client-ca\") pod \"f710888e-8c73-4d02-8ab4-f530b2562d8e\" (UID: \"f710888e-8c73-4d02-8ab4-f530b2562d8e\") " Jan 26 17:04:31 crc kubenswrapper[4856]: I0126 17:04:31.122854 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7x9xc\" (UniqueName: \"kubernetes.io/projected/f710888e-8c73-4d02-8ab4-f530b2562d8e-kube-api-access-7x9xc\") pod \"f710888e-8c73-4d02-8ab4-f530b2562d8e\" (UID: \"f710888e-8c73-4d02-8ab4-f530b2562d8e\") " Jan 26 17:04:31 crc kubenswrapper[4856]: I0126 17:04:31.123031 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brs28\" (UniqueName: \"kubernetes.io/projected/6651085c-1bee-4d12-baf3-469422e5d913-kube-api-access-brs28\") pod \"controller-manager-5674dc9874-8hwcb\" (UID: \"6651085c-1bee-4d12-baf3-469422e5d913\") " pod="openshift-controller-manager/controller-manager-5674dc9874-8hwcb" Jan 26 17:04:31 crc kubenswrapper[4856]: I0126 17:04:31.123648 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6651085c-1bee-4d12-baf3-469422e5d913-config\") pod \"controller-manager-5674dc9874-8hwcb\" (UID: \"6651085c-1bee-4d12-baf3-469422e5d913\") " pod="openshift-controller-manager/controller-manager-5674dc9874-8hwcb" Jan 26 17:04:31 crc kubenswrapper[4856]: I0126 17:04:31.123672 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6651085c-1bee-4d12-baf3-469422e5d913-serving-cert\") pod \"controller-manager-5674dc9874-8hwcb\" (UID: \"6651085c-1bee-4d12-baf3-469422e5d913\") " pod="openshift-controller-manager/controller-manager-5674dc9874-8hwcb" Jan 26 17:04:31 crc kubenswrapper[4856]: I0126 17:04:31.123672 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f710888e-8c73-4d02-8ab4-f530b2562d8e-config" (OuterVolumeSpecName: "config") pod "f710888e-8c73-4d02-8ab4-f530b2562d8e" (UID: "f710888e-8c73-4d02-8ab4-f530b2562d8e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:04:31 crc kubenswrapper[4856]: I0126 17:04:31.123674 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f710888e-8c73-4d02-8ab4-f530b2562d8e-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "f710888e-8c73-4d02-8ab4-f530b2562d8e" (UID: "f710888e-8c73-4d02-8ab4-f530b2562d8e"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:04:31 crc kubenswrapper[4856]: I0126 17:04:31.123865 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6651085c-1bee-4d12-baf3-469422e5d913-client-ca\") pod \"controller-manager-5674dc9874-8hwcb\" (UID: \"6651085c-1bee-4d12-baf3-469422e5d913\") " pod="openshift-controller-manager/controller-manager-5674dc9874-8hwcb" Jan 26 17:04:31 crc kubenswrapper[4856]: I0126 17:04:31.123919 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6651085c-1bee-4d12-baf3-469422e5d913-proxy-ca-bundles\") pod \"controller-manager-5674dc9874-8hwcb\" (UID: \"6651085c-1bee-4d12-baf3-469422e5d913\") " pod="openshift-controller-manager/controller-manager-5674dc9874-8hwcb" Jan 26 17:04:31 crc kubenswrapper[4856]: I0126 17:04:31.123900 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f710888e-8c73-4d02-8ab4-f530b2562d8e-client-ca" (OuterVolumeSpecName: "client-ca") pod "f710888e-8c73-4d02-8ab4-f530b2562d8e" (UID: "f710888e-8c73-4d02-8ab4-f530b2562d8e"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:04:31 crc kubenswrapper[4856]: I0126 17:04:31.124100 4856 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f710888e-8c73-4d02-8ab4-f530b2562d8e-config\") on node \"crc\" DevicePath \"\"" Jan 26 17:04:31 crc kubenswrapper[4856]: I0126 17:04:31.124266 4856 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f710888e-8c73-4d02-8ab4-f530b2562d8e-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 26 17:04:31 crc kubenswrapper[4856]: I0126 17:04:31.128985 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f710888e-8c73-4d02-8ab4-f530b2562d8e-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f710888e-8c73-4d02-8ab4-f530b2562d8e" (UID: "f710888e-8c73-4d02-8ab4-f530b2562d8e"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:04:31 crc kubenswrapper[4856]: I0126 17:04:31.132845 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f710888e-8c73-4d02-8ab4-f530b2562d8e-kube-api-access-7x9xc" (OuterVolumeSpecName: "kube-api-access-7x9xc") pod "f710888e-8c73-4d02-8ab4-f530b2562d8e" (UID: "f710888e-8c73-4d02-8ab4-f530b2562d8e"). InnerVolumeSpecName "kube-api-access-7x9xc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:04:31 crc kubenswrapper[4856]: I0126 17:04:31.225580 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6651085c-1bee-4d12-baf3-469422e5d913-config\") pod \"controller-manager-5674dc9874-8hwcb\" (UID: \"6651085c-1bee-4d12-baf3-469422e5d913\") " pod="openshift-controller-manager/controller-manager-5674dc9874-8hwcb" Jan 26 17:04:31 crc kubenswrapper[4856]: I0126 17:04:31.225628 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6651085c-1bee-4d12-baf3-469422e5d913-serving-cert\") pod \"controller-manager-5674dc9874-8hwcb\" (UID: \"6651085c-1bee-4d12-baf3-469422e5d913\") " pod="openshift-controller-manager/controller-manager-5674dc9874-8hwcb" Jan 26 17:04:31 crc kubenswrapper[4856]: I0126 17:04:31.225669 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6651085c-1bee-4d12-baf3-469422e5d913-client-ca\") pod \"controller-manager-5674dc9874-8hwcb\" (UID: \"6651085c-1bee-4d12-baf3-469422e5d913\") " pod="openshift-controller-manager/controller-manager-5674dc9874-8hwcb" Jan 26 17:04:31 crc kubenswrapper[4856]: I0126 17:04:31.225728 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6651085c-1bee-4d12-baf3-469422e5d913-proxy-ca-bundles\") pod \"controller-manager-5674dc9874-8hwcb\" (UID: \"6651085c-1bee-4d12-baf3-469422e5d913\") " pod="openshift-controller-manager/controller-manager-5674dc9874-8hwcb" Jan 26 17:04:31 crc kubenswrapper[4856]: I0126 17:04:31.225759 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brs28\" (UniqueName: \"kubernetes.io/projected/6651085c-1bee-4d12-baf3-469422e5d913-kube-api-access-brs28\") pod \"controller-manager-5674dc9874-8hwcb\" (UID: \"6651085c-1bee-4d12-baf3-469422e5d913\") " pod="openshift-controller-manager/controller-manager-5674dc9874-8hwcb" Jan 26 17:04:31 crc kubenswrapper[4856]: I0126 17:04:31.225832 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7x9xc\" (UniqueName: \"kubernetes.io/projected/f710888e-8c73-4d02-8ab4-f530b2562d8e-kube-api-access-7x9xc\") on node \"crc\" DevicePath \"\"" Jan 26 17:04:31 crc kubenswrapper[4856]: I0126 17:04:31.225849 4856 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f710888e-8c73-4d02-8ab4-f530b2562d8e-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 17:04:31 crc kubenswrapper[4856]: I0126 17:04:31.225863 4856 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f710888e-8c73-4d02-8ab4-f530b2562d8e-client-ca\") on node \"crc\" DevicePath \"\"" Jan 26 17:04:31 crc kubenswrapper[4856]: I0126 17:04:31.227251 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6651085c-1bee-4d12-baf3-469422e5d913-client-ca\") pod \"controller-manager-5674dc9874-8hwcb\" (UID: \"6651085c-1bee-4d12-baf3-469422e5d913\") " pod="openshift-controller-manager/controller-manager-5674dc9874-8hwcb" Jan 26 17:04:31 crc kubenswrapper[4856]: I0126 17:04:31.227249 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6651085c-1bee-4d12-baf3-469422e5d913-proxy-ca-bundles\") pod \"controller-manager-5674dc9874-8hwcb\" (UID: \"6651085c-1bee-4d12-baf3-469422e5d913\") " pod="openshift-controller-manager/controller-manager-5674dc9874-8hwcb" Jan 26 17:04:31 crc kubenswrapper[4856]: I0126 17:04:31.227427 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6651085c-1bee-4d12-baf3-469422e5d913-config\") pod \"controller-manager-5674dc9874-8hwcb\" (UID: \"6651085c-1bee-4d12-baf3-469422e5d913\") " pod="openshift-controller-manager/controller-manager-5674dc9874-8hwcb" Jan 26 17:04:31 crc kubenswrapper[4856]: I0126 17:04:31.232564 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6651085c-1bee-4d12-baf3-469422e5d913-serving-cert\") pod \"controller-manager-5674dc9874-8hwcb\" (UID: \"6651085c-1bee-4d12-baf3-469422e5d913\") " pod="openshift-controller-manager/controller-manager-5674dc9874-8hwcb" Jan 26 17:04:31 crc kubenswrapper[4856]: I0126 17:04:31.241466 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brs28\" (UniqueName: \"kubernetes.io/projected/6651085c-1bee-4d12-baf3-469422e5d913-kube-api-access-brs28\") pod \"controller-manager-5674dc9874-8hwcb\" (UID: \"6651085c-1bee-4d12-baf3-469422e5d913\") " pod="openshift-controller-manager/controller-manager-5674dc9874-8hwcb" Jan 26 17:04:31 crc kubenswrapper[4856]: I0126 17:04:31.345253 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5674dc9874-8hwcb" Jan 26 17:04:31 crc kubenswrapper[4856]: I0126 17:04:31.409372 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-787548c685-7dskp"] Jan 26 17:04:31 crc kubenswrapper[4856]: I0126 17:04:31.414749 4856 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-787548c685-7dskp"] Jan 26 17:04:31 crc kubenswrapper[4856]: I0126 17:04:31.766252 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5674dc9874-8hwcb"] Jan 26 17:04:31 crc kubenswrapper[4856]: W0126 17:04:31.773277 4856 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6651085c_1bee_4d12_baf3_469422e5d913.slice/crio-325d697f9bb3dc3b8b8d0844c09109968aa162bddb4a735c7737f413cf361e04 WatchSource:0}: Error finding container 325d697f9bb3dc3b8b8d0844c09109968aa162bddb4a735c7737f413cf361e04: Status 404 returned error can't find the container with id 325d697f9bb3dc3b8b8d0844c09109968aa162bddb4a735c7737f413cf361e04 Jan 26 17:04:32 crc kubenswrapper[4856]: I0126 17:04:32.075286 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5674dc9874-8hwcb" event={"ID":"6651085c-1bee-4d12-baf3-469422e5d913","Type":"ContainerStarted","Data":"873dac22c2f5e8d3028987af3886aaebbed0d2fee645d66d41add1b55db0fa4c"} Jan 26 17:04:32 crc kubenswrapper[4856]: I0126 17:04:32.075698 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5674dc9874-8hwcb" event={"ID":"6651085c-1bee-4d12-baf3-469422e5d913","Type":"ContainerStarted","Data":"325d697f9bb3dc3b8b8d0844c09109968aa162bddb4a735c7737f413cf361e04"} Jan 26 17:04:32 crc kubenswrapper[4856]: I0126 17:04:32.075723 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5674dc9874-8hwcb" Jan 26 17:04:32 crc kubenswrapper[4856]: I0126 17:04:32.081995 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5674dc9874-8hwcb" Jan 26 17:04:32 crc kubenswrapper[4856]: I0126 17:04:32.099368 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5674dc9874-8hwcb" podStartSLOduration=3.099325577 podStartE2EDuration="3.099325577s" podCreationTimestamp="2026-01-26 17:04:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:04:32.092142273 +0000 UTC m=+368.045396274" watchObservedRunningTime="2026-01-26 17:04:32.099325577 +0000 UTC m=+368.052579558" Jan 26 17:04:33 crc kubenswrapper[4856]: I0126 17:04:33.404459 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f710888e-8c73-4d02-8ab4-f530b2562d8e" path="/var/lib/kubelet/pods/f710888e-8c73-4d02-8ab4-f530b2562d8e/volumes" Jan 26 17:04:56 crc kubenswrapper[4856]: I0126 17:04:56.939687 4856 patch_prober.go:28] interesting pod/machine-config-daemon-xm9cq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:04:56 crc kubenswrapper[4856]: I0126 17:04:56.940311 4856 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:05:25 crc kubenswrapper[4856]: I0126 17:05:25.627051 4856 scope.go:117] "RemoveContainer" containerID="59ba0f131eee14df389391e46bc772894e49e976c3663df5fb3bc98b3cb4a3d6" Jan 26 17:05:25 crc kubenswrapper[4856]: I0126 17:05:25.646487 4856 scope.go:117] "RemoveContainer" containerID="2ffc4b80383322e2c628a33ac37f15fd77c7650ca108c022f97bb48aad023462" Jan 26 17:05:25 crc kubenswrapper[4856]: I0126 17:05:25.665620 4856 scope.go:117] "RemoveContainer" containerID="40d432159a07ba20bf95f058bf8a597d67cbd8d852519bed035694f8ba3d8ec4" Jan 26 17:05:25 crc kubenswrapper[4856]: I0126 17:05:25.696964 4856 scope.go:117] "RemoveContainer" containerID="89975b8f9428f81ab5d3fb48ced5dd9c837bea2feea3b89f5f7ff8d7d5d15b3e" Jan 26 17:05:25 crc kubenswrapper[4856]: I0126 17:05:25.715696 4856 scope.go:117] "RemoveContainer" containerID="b03a01b651d9f66da4dd1f6e0d29ad97c0d6ae46b644c3d997d8dc99476706df" Jan 26 17:05:25 crc kubenswrapper[4856]: I0126 17:05:25.732128 4856 scope.go:117] "RemoveContainer" containerID="f2daa3f13d37b7ae19dfca406b6b5cfdcd4f211287f7d410193f2fee36a24553" Jan 26 17:05:26 crc kubenswrapper[4856]: I0126 17:05:26.939026 4856 patch_prober.go:28] interesting pod/machine-config-daemon-xm9cq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:05:26 crc kubenswrapper[4856]: I0126 17:05:26.941394 4856 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:05:56 crc kubenswrapper[4856]: I0126 17:05:56.939685 4856 patch_prober.go:28] interesting pod/machine-config-daemon-xm9cq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:05:56 crc kubenswrapper[4856]: I0126 17:05:56.940231 4856 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:05:56 crc kubenswrapper[4856]: I0126 17:05:56.940350 4856 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" Jan 26 17:05:56 crc kubenswrapper[4856]: I0126 17:05:56.941255 4856 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9758bfdfd1807e791935ac7ec93246863e5867351e35d27ffaff68ae79110e9c"} pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 26 17:05:56 crc kubenswrapper[4856]: I0126 17:05:56.941338 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" containerName="machine-config-daemon" containerID="cri-o://9758bfdfd1807e791935ac7ec93246863e5867351e35d27ffaff68ae79110e9c" gracePeriod=600 Jan 26 17:05:57 crc kubenswrapper[4856]: I0126 17:05:57.609018 4856 generic.go:334] "Generic (PLEG): container finished" podID="63c75ede-5170-4db0-811b-5217ef8d72b3" containerID="9758bfdfd1807e791935ac7ec93246863e5867351e35d27ffaff68ae79110e9c" exitCode=0 Jan 26 17:05:57 crc kubenswrapper[4856]: I0126 17:05:57.609088 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" event={"ID":"63c75ede-5170-4db0-811b-5217ef8d72b3","Type":"ContainerDied","Data":"9758bfdfd1807e791935ac7ec93246863e5867351e35d27ffaff68ae79110e9c"} Jan 26 17:05:57 crc kubenswrapper[4856]: I0126 17:05:57.609478 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" event={"ID":"63c75ede-5170-4db0-811b-5217ef8d72b3","Type":"ContainerStarted","Data":"fe42c0299ac9f35a2260caaf7226f7e2161da013442117dab0d25a7c69c46115"} Jan 26 17:05:57 crc kubenswrapper[4856]: I0126 17:05:57.609506 4856 scope.go:117] "RemoveContainer" containerID="54ca3fa13d9e8d442efa93b44a870369f7df3fe7562d77b98528f5c19a751f18" Jan 26 17:07:25 crc kubenswrapper[4856]: I0126 17:07:25.784371 4856 scope.go:117] "RemoveContainer" containerID="f96de8f882682ea8e5a30970c1ce8d34c4b60cb434e13968e3bd6879b62b071b" Jan 26 17:08:26 crc kubenswrapper[4856]: I0126 17:08:26.938550 4856 patch_prober.go:28] interesting pod/machine-config-daemon-xm9cq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:08:26 crc kubenswrapper[4856]: I0126 17:08:26.939789 4856 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:08:56 crc kubenswrapper[4856]: I0126 17:08:56.939167 4856 patch_prober.go:28] interesting pod/machine-config-daemon-xm9cq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:08:56 crc kubenswrapper[4856]: I0126 17:08:56.939502 4856 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:09:17 crc kubenswrapper[4856]: I0126 17:09:17.926299 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pxh94"] Jan 26 17:09:17 crc kubenswrapper[4856]: I0126 17:09:17.929212 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="sbdb" containerID="cri-o://11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a" gracePeriod=30 Jan 26 17:09:17 crc kubenswrapper[4856]: I0126 17:09:17.929247 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3" gracePeriod=30 Jan 26 17:09:17 crc kubenswrapper[4856]: I0126 17:09:17.929211 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="nbdb" containerID="cri-o://4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7" gracePeriod=30 Jan 26 17:09:17 crc kubenswrapper[4856]: I0126 17:09:17.929320 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="northd" containerID="cri-o://b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde" gracePeriod=30 Jan 26 17:09:17 crc kubenswrapper[4856]: I0126 17:09:17.929418 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="ovn-acl-logging" containerID="cri-o://83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc" gracePeriod=30 Jan 26 17:09:17 crc kubenswrapper[4856]: I0126 17:09:17.929412 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="kube-rbac-proxy-node" containerID="cri-o://e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1" gracePeriod=30 Jan 26 17:09:17 crc kubenswrapper[4856]: I0126 17:09:17.929130 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="ovn-controller" containerID="cri-o://25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b" gracePeriod=30 Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.000696 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="ovnkube-controller" containerID="cri-o://984517fa62d3e73cf4b33b7c4a101f8221c940f66938120373b7d41cac9c5e5c" gracePeriod=30 Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.228596 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pxh94_ab5b6f50-172b-4535-a0f9-5d103bcab4e7/ovnkube-controller/3.log" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.232373 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pxh94_ab5b6f50-172b-4535-a0f9-5d103bcab4e7/ovn-acl-logging/0.log" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.233280 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pxh94_ab5b6f50-172b-4535-a0f9-5d103bcab4e7/ovn-controller/0.log" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.235919 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.298877 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6t25x"] Jan 26 17:09:18 crc kubenswrapper[4856]: E0126 17:09:18.299204 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="ovnkube-controller" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.299225 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="ovnkube-controller" Jan 26 17:09:18 crc kubenswrapper[4856]: E0126 17:09:18.299234 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="ovnkube-controller" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.299239 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="ovnkube-controller" Jan 26 17:09:18 crc kubenswrapper[4856]: E0126 17:09:18.299247 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="kube-rbac-proxy-node" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.299252 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="kube-rbac-proxy-node" Jan 26 17:09:18 crc kubenswrapper[4856]: E0126 17:09:18.299261 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="sbdb" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.299266 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="sbdb" Jan 26 17:09:18 crc kubenswrapper[4856]: E0126 17:09:18.299279 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="ovn-acl-logging" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.299285 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="ovn-acl-logging" Jan 26 17:09:18 crc kubenswrapper[4856]: E0126 17:09:18.299293 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="northd" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.299299 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="northd" Jan 26 17:09:18 crc kubenswrapper[4856]: E0126 17:09:18.299308 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="ovn-controller" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.299314 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="ovn-controller" Jan 26 17:09:18 crc kubenswrapper[4856]: E0126 17:09:18.299325 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="nbdb" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.299330 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="nbdb" Jan 26 17:09:18 crc kubenswrapper[4856]: E0126 17:09:18.299339 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="kubecfg-setup" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.299344 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="kubecfg-setup" Jan 26 17:09:18 crc kubenswrapper[4856]: E0126 17:09:18.299352 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="ovnkube-controller" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.299357 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="ovnkube-controller" Jan 26 17:09:18 crc kubenswrapper[4856]: E0126 17:09:18.299366 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="kube-rbac-proxy-ovn-metrics" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.299372 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="kube-rbac-proxy-ovn-metrics" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.299491 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="ovn-controller" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.299499 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="ovnkube-controller" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.299511 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="sbdb" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.299517 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="ovnkube-controller" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.299526 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="ovnkube-controller" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.299551 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="kube-rbac-proxy-node" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.299560 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="ovn-acl-logging" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.299568 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="kube-rbac-proxy-ovn-metrics" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.299578 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="nbdb" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.299585 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="northd" Jan 26 17:09:18 crc kubenswrapper[4856]: E0126 17:09:18.299670 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="ovnkube-controller" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.299685 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="ovnkube-controller" Jan 26 17:09:18 crc kubenswrapper[4856]: E0126 17:09:18.299692 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="ovnkube-controller" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.299698 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="ovnkube-controller" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.299787 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="ovnkube-controller" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.299795 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerName="ovnkube-controller" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.301448 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.422913 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-run-ovn\") pod \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.422978 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-host-run-netns\") pod \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.423039 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-host-cni-netd\") pod \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.423083 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "ab5b6f50-172b-4535-a0f9-5d103bcab4e7" (UID: "ab5b6f50-172b-4535-a0f9-5d103bcab4e7"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.423098 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "ab5b6f50-172b-4535-a0f9-5d103bcab4e7" (UID: "ab5b6f50-172b-4535-a0f9-5d103bcab4e7"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.423118 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "ab5b6f50-172b-4535-a0f9-5d103bcab4e7" (UID: "ab5b6f50-172b-4535-a0f9-5d103bcab4e7"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.423770 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-host-cni-bin\") pod \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.423839 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-ovn-node-metrics-cert\") pod \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.423871 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-systemd-units\") pod \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.423918 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-host-kubelet\") pod \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.423944 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-etc-openvswitch\") pod \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.423971 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-run-openvswitch\") pod \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.424015 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-run-systemd\") pod \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.424051 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-host-run-ovn-kubernetes\") pod \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.424086 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-env-overrides\") pod \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.424114 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-node-log\") pod \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.424147 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-ovnkube-config\") pod \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.424174 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.424217 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-ovnkube-script-lib\") pod \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.424249 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-var-lib-openvswitch\") pod \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.424280 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9kdbz\" (UniqueName: \"kubernetes.io/projected/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-kube-api-access-9kdbz\") pod \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.424311 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-log-socket\") pod \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.424338 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-host-slash\") pod \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\" (UID: \"ab5b6f50-172b-4535-a0f9-5d103bcab4e7\") " Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.424010 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "ab5b6f50-172b-4535-a0f9-5d103bcab4e7" (UID: "ab5b6f50-172b-4535-a0f9-5d103bcab4e7"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.425630 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-node-log" (OuterVolumeSpecName: "node-log") pod "ab5b6f50-172b-4535-a0f9-5d103bcab4e7" (UID: "ab5b6f50-172b-4535-a0f9-5d103bcab4e7"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.424034 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "ab5b6f50-172b-4535-a0f9-5d103bcab4e7" (UID: "ab5b6f50-172b-4535-a0f9-5d103bcab4e7"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.425277 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "ab5b6f50-172b-4535-a0f9-5d103bcab4e7" (UID: "ab5b6f50-172b-4535-a0f9-5d103bcab4e7"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.425340 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-log-socket" (OuterVolumeSpecName: "log-socket") pod "ab5b6f50-172b-4535-a0f9-5d103bcab4e7" (UID: "ab5b6f50-172b-4535-a0f9-5d103bcab4e7"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.425386 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "ab5b6f50-172b-4535-a0f9-5d103bcab4e7" (UID: "ab5b6f50-172b-4535-a0f9-5d103bcab4e7"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.425685 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "ab5b6f50-172b-4535-a0f9-5d103bcab4e7" (UID: "ab5b6f50-172b-4535-a0f9-5d103bcab4e7"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.425393 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "ab5b6f50-172b-4535-a0f9-5d103bcab4e7" (UID: "ab5b6f50-172b-4535-a0f9-5d103bcab4e7"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.425415 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "ab5b6f50-172b-4535-a0f9-5d103bcab4e7" (UID: "ab5b6f50-172b-4535-a0f9-5d103bcab4e7"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.425601 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "ab5b6f50-172b-4535-a0f9-5d103bcab4e7" (UID: "ab5b6f50-172b-4535-a0f9-5d103bcab4e7"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.425634 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "ab5b6f50-172b-4535-a0f9-5d103bcab4e7" (UID: "ab5b6f50-172b-4535-a0f9-5d103bcab4e7"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.425772 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "ab5b6f50-172b-4535-a0f9-5d103bcab4e7" (UID: "ab5b6f50-172b-4535-a0f9-5d103bcab4e7"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.425920 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "ab5b6f50-172b-4535-a0f9-5d103bcab4e7" (UID: "ab5b6f50-172b-4535-a0f9-5d103bcab4e7"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.427089 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d70f72f7-ff5e-4906-8622-9cddfe769d55-ovnkube-config\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.427231 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d70f72f7-ff5e-4906-8622-9cddfe769d55-env-overrides\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.427279 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-host-kubelet\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.427310 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-host-run-ovn-kubernetes\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.427353 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnrjd\" (UniqueName: \"kubernetes.io/projected/d70f72f7-ff5e-4906-8622-9cddfe769d55-kube-api-access-cnrjd\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.427378 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-host-run-netns\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.427405 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d70f72f7-ff5e-4906-8622-9cddfe769d55-ovnkube-script-lib\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.427435 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-var-lib-openvswitch\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.427470 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-etc-openvswitch\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.427756 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-run-openvswitch\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.427796 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-run-ovn\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.427857 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d70f72f7-ff5e-4906-8622-9cddfe769d55-ovn-node-metrics-cert\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.427900 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-run-systemd\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.427929 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-log-socket\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.427956 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-host-cni-bin\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.427995 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-node-log\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.428132 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-host-cni-netd\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.428173 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.428561 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-host-slash\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.428674 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-systemd-units\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.428973 4856 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.428998 4856 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.429017 4856 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-log-socket\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.429033 4856 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-run-ovn\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.429051 4856 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-host-run-netns\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.429063 4856 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-host-cni-netd\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.429075 4856 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-host-cni-bin\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.429088 4856 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-systemd-units\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.429137 4856 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.429151 4856 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-host-kubelet\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.429165 4856 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-run-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.429183 4856 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.429199 4856 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.429211 4856 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-node-log\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.429223 4856 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.429240 4856 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.429974 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "ab5b6f50-172b-4535-a0f9-5d103bcab4e7" (UID: "ab5b6f50-172b-4535-a0f9-5d103bcab4e7"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.430326 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-kube-api-access-9kdbz" (OuterVolumeSpecName: "kube-api-access-9kdbz") pod "ab5b6f50-172b-4535-a0f9-5d103bcab4e7" (UID: "ab5b6f50-172b-4535-a0f9-5d103bcab4e7"). InnerVolumeSpecName "kube-api-access-9kdbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.438070 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-host-slash" (OuterVolumeSpecName: "host-slash") pod "ab5b6f50-172b-4535-a0f9-5d103bcab4e7" (UID: "ab5b6f50-172b-4535-a0f9-5d103bcab4e7"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.440137 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "ab5b6f50-172b-4535-a0f9-5d103bcab4e7" (UID: "ab5b6f50-172b-4535-a0f9-5d103bcab4e7"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.530008 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-node-log\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.530083 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.530113 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-host-cni-netd\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.530141 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-host-slash\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.530162 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-systemd-units\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.530194 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d70f72f7-ff5e-4906-8622-9cddfe769d55-ovnkube-config\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.530185 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-node-log\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.530217 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d70f72f7-ff5e-4906-8622-9cddfe769d55-env-overrides\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.530236 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-host-cni-netd\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.530274 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-host-kubelet\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.530277 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-systemd-units\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.530245 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-host-kubelet\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.530290 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.530351 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-host-run-ovn-kubernetes\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.530385 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnrjd\" (UniqueName: \"kubernetes.io/projected/d70f72f7-ff5e-4906-8622-9cddfe769d55-kube-api-access-cnrjd\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.530406 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d70f72f7-ff5e-4906-8622-9cddfe769d55-ovnkube-script-lib\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.530434 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-host-run-netns\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.530455 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-var-lib-openvswitch\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.530459 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-host-run-ovn-kubernetes\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.530493 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-etc-openvswitch\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.530516 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-run-openvswitch\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.530518 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-host-run-netns\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.530576 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-run-ovn\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.530647 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d70f72f7-ff5e-4906-8622-9cddfe769d55-ovn-node-metrics-cert\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.530683 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-run-systemd\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.530423 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-host-slash\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.530734 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-etc-openvswitch\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.530780 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-run-openvswitch\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.530797 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-run-ovn\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.530810 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-var-lib-openvswitch\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.530837 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-log-socket\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.530847 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-run-systemd\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.530870 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-host-cni-bin\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.530869 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-log-socket\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.530922 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d70f72f7-ff5e-4906-8622-9cddfe769d55-env-overrides\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.530983 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d70f72f7-ff5e-4906-8622-9cddfe769d55-host-cni-bin\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.531042 4856 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.531063 4856 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-run-systemd\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.531075 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9kdbz\" (UniqueName: \"kubernetes.io/projected/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-kube-api-access-9kdbz\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.531084 4856 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ab5b6f50-172b-4535-a0f9-5d103bcab4e7-host-slash\") on node \"crc\" DevicePath \"\"" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.531422 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d70f72f7-ff5e-4906-8622-9cddfe769d55-ovnkube-config\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.532294 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d70f72f7-ff5e-4906-8622-9cddfe769d55-ovnkube-script-lib\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.533872 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d70f72f7-ff5e-4906-8622-9cddfe769d55-ovn-node-metrics-cert\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.549066 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnrjd\" (UniqueName: \"kubernetes.io/projected/d70f72f7-ff5e-4906-8622-9cddfe769d55-kube-api-access-cnrjd\") pod \"ovnkube-node-6t25x\" (UID: \"d70f72f7-ff5e-4906-8622-9cddfe769d55\") " pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.617865 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.968583 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pxh94_ab5b6f50-172b-4535-a0f9-5d103bcab4e7/ovnkube-controller/3.log" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.971729 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pxh94_ab5b6f50-172b-4535-a0f9-5d103bcab4e7/ovn-acl-logging/0.log" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.972363 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pxh94_ab5b6f50-172b-4535-a0f9-5d103bcab4e7/ovn-controller/0.log" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.972890 4856 generic.go:334] "Generic (PLEG): container finished" podID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerID="984517fa62d3e73cf4b33b7c4a101f8221c940f66938120373b7d41cac9c5e5c" exitCode=0 Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.972917 4856 generic.go:334] "Generic (PLEG): container finished" podID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerID="11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a" exitCode=0 Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.972925 4856 generic.go:334] "Generic (PLEG): container finished" podID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerID="4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7" exitCode=0 Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.972948 4856 generic.go:334] "Generic (PLEG): container finished" podID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerID="b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde" exitCode=0 Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.972955 4856 generic.go:334] "Generic (PLEG): container finished" podID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerID="7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3" exitCode=0 Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.972962 4856 generic.go:334] "Generic (PLEG): container finished" podID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerID="e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1" exitCode=0 Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.972969 4856 generic.go:334] "Generic (PLEG): container finished" podID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerID="83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc" exitCode=143 Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.972975 4856 generic.go:334] "Generic (PLEG): container finished" podID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" containerID="25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b" exitCode=143 Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.972973 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" event={"ID":"ab5b6f50-172b-4535-a0f9-5d103bcab4e7","Type":"ContainerDied","Data":"984517fa62d3e73cf4b33b7c4a101f8221c940f66938120373b7d41cac9c5e5c"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973005 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973036 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" event={"ID":"ab5b6f50-172b-4535-a0f9-5d103bcab4e7","Type":"ContainerDied","Data":"11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973051 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" event={"ID":"ab5b6f50-172b-4535-a0f9-5d103bcab4e7","Type":"ContainerDied","Data":"4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973062 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" event={"ID":"ab5b6f50-172b-4535-a0f9-5d103bcab4e7","Type":"ContainerDied","Data":"b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973073 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" event={"ID":"ab5b6f50-172b-4535-a0f9-5d103bcab4e7","Type":"ContainerDied","Data":"7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973082 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" event={"ID":"ab5b6f50-172b-4535-a0f9-5d103bcab4e7","Type":"ContainerDied","Data":"e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973103 4856 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"203d756903498bed1c57a3e87a95f4b24f808514567a99505ba2f3cfec468cb6"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973118 4856 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973124 4856 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973126 4856 scope.go:117] "RemoveContainer" containerID="984517fa62d3e73cf4b33b7c4a101f8221c940f66938120373b7d41cac9c5e5c" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973130 4856 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973214 4856 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973226 4856 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973233 4856 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973238 4856 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973244 4856 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973277 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" event={"ID":"ab5b6f50-172b-4535-a0f9-5d103bcab4e7","Type":"ContainerDied","Data":"83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973294 4856 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"984517fa62d3e73cf4b33b7c4a101f8221c940f66938120373b7d41cac9c5e5c"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973300 4856 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"203d756903498bed1c57a3e87a95f4b24f808514567a99505ba2f3cfec468cb6"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973305 4856 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973311 4856 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973316 4856 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973321 4856 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973342 4856 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973347 4856 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973353 4856 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973359 4856 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973368 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" event={"ID":"ab5b6f50-172b-4535-a0f9-5d103bcab4e7","Type":"ContainerDied","Data":"25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973378 4856 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"984517fa62d3e73cf4b33b7c4a101f8221c940f66938120373b7d41cac9c5e5c"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973386 4856 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"203d756903498bed1c57a3e87a95f4b24f808514567a99505ba2f3cfec468cb6"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973393 4856 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973399 4856 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973420 4856 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973426 4856 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973431 4856 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973436 4856 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973441 4856 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973446 4856 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973453 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pxh94" event={"ID":"ab5b6f50-172b-4535-a0f9-5d103bcab4e7","Type":"ContainerDied","Data":"a1b2fe845f0957cc37219c78a754b5c2b9acc25bf2ef8f7083ca734c4c5c68b9"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973460 4856 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"984517fa62d3e73cf4b33b7c4a101f8221c940f66938120373b7d41cac9c5e5c"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973466 4856 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"203d756903498bed1c57a3e87a95f4b24f808514567a99505ba2f3cfec468cb6"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973472 4856 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973477 4856 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973496 4856 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973502 4856 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973507 4856 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973512 4856 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973517 4856 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.973546 4856 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.975905 4856 generic.go:334] "Generic (PLEG): container finished" podID="d70f72f7-ff5e-4906-8622-9cddfe769d55" containerID="77c4fd70275b917cb0f73727113a417c575c523f2153a6e00e93ec081a5c0141" exitCode=0 Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.975927 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" event={"ID":"d70f72f7-ff5e-4906-8622-9cddfe769d55","Type":"ContainerDied","Data":"77c4fd70275b917cb0f73727113a417c575c523f2153a6e00e93ec081a5c0141"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.975948 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" event={"ID":"d70f72f7-ff5e-4906-8622-9cddfe769d55","Type":"ContainerStarted","Data":"44ee9c730d2fa91ea1f9b547ba330dbc15901ea95b7027620fc8366ea50bb691"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.979646 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-rq622_7a742e7b-c420-46e3-9e96-e9c744af6124/kube-multus/2.log" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.980580 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-rq622_7a742e7b-c420-46e3-9e96-e9c744af6124/kube-multus/1.log" Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.980622 4856 generic.go:334] "Generic (PLEG): container finished" podID="7a742e7b-c420-46e3-9e96-e9c744af6124" containerID="ddec0dbea657c6160cfdfd78886d5ae335dab8b667b0e0e3813dffa86a2ae2dc" exitCode=2 Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.980654 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-rq622" event={"ID":"7a742e7b-c420-46e3-9e96-e9c744af6124","Type":"ContainerDied","Data":"ddec0dbea657c6160cfdfd78886d5ae335dab8b667b0e0e3813dffa86a2ae2dc"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.980676 4856 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"afeb20035224feeab28a92ac77b43a24e653e49c56a25590a9861019a2b7a8ff"} Jan 26 17:09:18 crc kubenswrapper[4856]: I0126 17:09:18.981090 4856 scope.go:117] "RemoveContainer" containerID="ddec0dbea657c6160cfdfd78886d5ae335dab8b667b0e0e3813dffa86a2ae2dc" Jan 26 17:09:18 crc kubenswrapper[4856]: E0126 17:09:18.981343 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-rq622_openshift-multus(7a742e7b-c420-46e3-9e96-e9c744af6124)\"" pod="openshift-multus/multus-rq622" podUID="7a742e7b-c420-46e3-9e96-e9c744af6124" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.002550 4856 scope.go:117] "RemoveContainer" containerID="203d756903498bed1c57a3e87a95f4b24f808514567a99505ba2f3cfec468cb6" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.025899 4856 scope.go:117] "RemoveContainer" containerID="11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.046483 4856 scope.go:117] "RemoveContainer" containerID="4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.046628 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pxh94"] Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.051179 4856 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pxh94"] Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.087009 4856 scope.go:117] "RemoveContainer" containerID="b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.099405 4856 scope.go:117] "RemoveContainer" containerID="7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.113323 4856 scope.go:117] "RemoveContainer" containerID="e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.126899 4856 scope.go:117] "RemoveContainer" containerID="83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.148402 4856 scope.go:117] "RemoveContainer" containerID="25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.167511 4856 scope.go:117] "RemoveContainer" containerID="d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.180249 4856 scope.go:117] "RemoveContainer" containerID="984517fa62d3e73cf4b33b7c4a101f8221c940f66938120373b7d41cac9c5e5c" Jan 26 17:09:19 crc kubenswrapper[4856]: E0126 17:09:19.180816 4856 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"984517fa62d3e73cf4b33b7c4a101f8221c940f66938120373b7d41cac9c5e5c\": container with ID starting with 984517fa62d3e73cf4b33b7c4a101f8221c940f66938120373b7d41cac9c5e5c not found: ID does not exist" containerID="984517fa62d3e73cf4b33b7c4a101f8221c940f66938120373b7d41cac9c5e5c" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.180854 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"984517fa62d3e73cf4b33b7c4a101f8221c940f66938120373b7d41cac9c5e5c"} err="failed to get container status \"984517fa62d3e73cf4b33b7c4a101f8221c940f66938120373b7d41cac9c5e5c\": rpc error: code = NotFound desc = could not find container \"984517fa62d3e73cf4b33b7c4a101f8221c940f66938120373b7d41cac9c5e5c\": container with ID starting with 984517fa62d3e73cf4b33b7c4a101f8221c940f66938120373b7d41cac9c5e5c not found: ID does not exist" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.180881 4856 scope.go:117] "RemoveContainer" containerID="203d756903498bed1c57a3e87a95f4b24f808514567a99505ba2f3cfec468cb6" Jan 26 17:09:19 crc kubenswrapper[4856]: E0126 17:09:19.181292 4856 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"203d756903498bed1c57a3e87a95f4b24f808514567a99505ba2f3cfec468cb6\": container with ID starting with 203d756903498bed1c57a3e87a95f4b24f808514567a99505ba2f3cfec468cb6 not found: ID does not exist" containerID="203d756903498bed1c57a3e87a95f4b24f808514567a99505ba2f3cfec468cb6" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.181316 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"203d756903498bed1c57a3e87a95f4b24f808514567a99505ba2f3cfec468cb6"} err="failed to get container status \"203d756903498bed1c57a3e87a95f4b24f808514567a99505ba2f3cfec468cb6\": rpc error: code = NotFound desc = could not find container \"203d756903498bed1c57a3e87a95f4b24f808514567a99505ba2f3cfec468cb6\": container with ID starting with 203d756903498bed1c57a3e87a95f4b24f808514567a99505ba2f3cfec468cb6 not found: ID does not exist" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.181336 4856 scope.go:117] "RemoveContainer" containerID="11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a" Jan 26 17:09:19 crc kubenswrapper[4856]: E0126 17:09:19.181631 4856 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a\": container with ID starting with 11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a not found: ID does not exist" containerID="11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.181725 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a"} err="failed to get container status \"11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a\": rpc error: code = NotFound desc = could not find container \"11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a\": container with ID starting with 11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a not found: ID does not exist" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.181790 4856 scope.go:117] "RemoveContainer" containerID="4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7" Jan 26 17:09:19 crc kubenswrapper[4856]: E0126 17:09:19.182202 4856 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7\": container with ID starting with 4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7 not found: ID does not exist" containerID="4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.182262 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7"} err="failed to get container status \"4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7\": rpc error: code = NotFound desc = could not find container \"4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7\": container with ID starting with 4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7 not found: ID does not exist" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.182303 4856 scope.go:117] "RemoveContainer" containerID="b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde" Jan 26 17:09:19 crc kubenswrapper[4856]: E0126 17:09:19.182726 4856 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde\": container with ID starting with b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde not found: ID does not exist" containerID="b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.182807 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde"} err="failed to get container status \"b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde\": rpc error: code = NotFound desc = could not find container \"b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde\": container with ID starting with b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde not found: ID does not exist" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.182888 4856 scope.go:117] "RemoveContainer" containerID="7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3" Jan 26 17:09:19 crc kubenswrapper[4856]: E0126 17:09:19.183264 4856 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3\": container with ID starting with 7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3 not found: ID does not exist" containerID="7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.183360 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3"} err="failed to get container status \"7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3\": rpc error: code = NotFound desc = could not find container \"7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3\": container with ID starting with 7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3 not found: ID does not exist" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.183425 4856 scope.go:117] "RemoveContainer" containerID="e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1" Jan 26 17:09:19 crc kubenswrapper[4856]: E0126 17:09:19.183854 4856 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1\": container with ID starting with e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1 not found: ID does not exist" containerID="e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.183964 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1"} err="failed to get container status \"e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1\": rpc error: code = NotFound desc = could not find container \"e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1\": container with ID starting with e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1 not found: ID does not exist" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.184041 4856 scope.go:117] "RemoveContainer" containerID="83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc" Jan 26 17:09:19 crc kubenswrapper[4856]: E0126 17:09:19.184417 4856 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc\": container with ID starting with 83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc not found: ID does not exist" containerID="83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.184440 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc"} err="failed to get container status \"83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc\": rpc error: code = NotFound desc = could not find container \"83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc\": container with ID starting with 83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc not found: ID does not exist" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.184457 4856 scope.go:117] "RemoveContainer" containerID="25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b" Jan 26 17:09:19 crc kubenswrapper[4856]: E0126 17:09:19.184722 4856 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b\": container with ID starting with 25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b not found: ID does not exist" containerID="25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.184748 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b"} err="failed to get container status \"25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b\": rpc error: code = NotFound desc = could not find container \"25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b\": container with ID starting with 25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b not found: ID does not exist" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.184778 4856 scope.go:117] "RemoveContainer" containerID="d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8" Jan 26 17:09:19 crc kubenswrapper[4856]: E0126 17:09:19.185019 4856 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\": container with ID starting with d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8 not found: ID does not exist" containerID="d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.185161 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8"} err="failed to get container status \"d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\": rpc error: code = NotFound desc = could not find container \"d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\": container with ID starting with d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8 not found: ID does not exist" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.185263 4856 scope.go:117] "RemoveContainer" containerID="984517fa62d3e73cf4b33b7c4a101f8221c940f66938120373b7d41cac9c5e5c" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.185694 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"984517fa62d3e73cf4b33b7c4a101f8221c940f66938120373b7d41cac9c5e5c"} err="failed to get container status \"984517fa62d3e73cf4b33b7c4a101f8221c940f66938120373b7d41cac9c5e5c\": rpc error: code = NotFound desc = could not find container \"984517fa62d3e73cf4b33b7c4a101f8221c940f66938120373b7d41cac9c5e5c\": container with ID starting with 984517fa62d3e73cf4b33b7c4a101f8221c940f66938120373b7d41cac9c5e5c not found: ID does not exist" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.185737 4856 scope.go:117] "RemoveContainer" containerID="203d756903498bed1c57a3e87a95f4b24f808514567a99505ba2f3cfec468cb6" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.186012 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"203d756903498bed1c57a3e87a95f4b24f808514567a99505ba2f3cfec468cb6"} err="failed to get container status \"203d756903498bed1c57a3e87a95f4b24f808514567a99505ba2f3cfec468cb6\": rpc error: code = NotFound desc = could not find container \"203d756903498bed1c57a3e87a95f4b24f808514567a99505ba2f3cfec468cb6\": container with ID starting with 203d756903498bed1c57a3e87a95f4b24f808514567a99505ba2f3cfec468cb6 not found: ID does not exist" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.186154 4856 scope.go:117] "RemoveContainer" containerID="11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.186569 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a"} err="failed to get container status \"11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a\": rpc error: code = NotFound desc = could not find container \"11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a\": container with ID starting with 11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a not found: ID does not exist" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.186590 4856 scope.go:117] "RemoveContainer" containerID="4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.187113 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7"} err="failed to get container status \"4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7\": rpc error: code = NotFound desc = could not find container \"4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7\": container with ID starting with 4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7 not found: ID does not exist" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.187228 4856 scope.go:117] "RemoveContainer" containerID="b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.187715 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde"} err="failed to get container status \"b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde\": rpc error: code = NotFound desc = could not find container \"b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde\": container with ID starting with b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde not found: ID does not exist" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.187809 4856 scope.go:117] "RemoveContainer" containerID="7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.188135 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3"} err="failed to get container status \"7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3\": rpc error: code = NotFound desc = could not find container \"7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3\": container with ID starting with 7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3 not found: ID does not exist" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.188168 4856 scope.go:117] "RemoveContainer" containerID="e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.188409 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1"} err="failed to get container status \"e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1\": rpc error: code = NotFound desc = could not find container \"e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1\": container with ID starting with e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1 not found: ID does not exist" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.188488 4856 scope.go:117] "RemoveContainer" containerID="83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.188823 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc"} err="failed to get container status \"83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc\": rpc error: code = NotFound desc = could not find container \"83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc\": container with ID starting with 83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc not found: ID does not exist" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.188850 4856 scope.go:117] "RemoveContainer" containerID="25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.189138 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b"} err="failed to get container status \"25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b\": rpc error: code = NotFound desc = could not find container \"25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b\": container with ID starting with 25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b not found: ID does not exist" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.189240 4856 scope.go:117] "RemoveContainer" containerID="d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.189542 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8"} err="failed to get container status \"d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\": rpc error: code = NotFound desc = could not find container \"d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\": container with ID starting with d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8 not found: ID does not exist" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.189564 4856 scope.go:117] "RemoveContainer" containerID="984517fa62d3e73cf4b33b7c4a101f8221c940f66938120373b7d41cac9c5e5c" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.189954 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"984517fa62d3e73cf4b33b7c4a101f8221c940f66938120373b7d41cac9c5e5c"} err="failed to get container status \"984517fa62d3e73cf4b33b7c4a101f8221c940f66938120373b7d41cac9c5e5c\": rpc error: code = NotFound desc = could not find container \"984517fa62d3e73cf4b33b7c4a101f8221c940f66938120373b7d41cac9c5e5c\": container with ID starting with 984517fa62d3e73cf4b33b7c4a101f8221c940f66938120373b7d41cac9c5e5c not found: ID does not exist" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.189984 4856 scope.go:117] "RemoveContainer" containerID="203d756903498bed1c57a3e87a95f4b24f808514567a99505ba2f3cfec468cb6" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.190294 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"203d756903498bed1c57a3e87a95f4b24f808514567a99505ba2f3cfec468cb6"} err="failed to get container status \"203d756903498bed1c57a3e87a95f4b24f808514567a99505ba2f3cfec468cb6\": rpc error: code = NotFound desc = could not find container \"203d756903498bed1c57a3e87a95f4b24f808514567a99505ba2f3cfec468cb6\": container with ID starting with 203d756903498bed1c57a3e87a95f4b24f808514567a99505ba2f3cfec468cb6 not found: ID does not exist" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.190322 4856 scope.go:117] "RemoveContainer" containerID="11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.190624 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a"} err="failed to get container status \"11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a\": rpc error: code = NotFound desc = could not find container \"11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a\": container with ID starting with 11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a not found: ID does not exist" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.190655 4856 scope.go:117] "RemoveContainer" containerID="4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.190885 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7"} err="failed to get container status \"4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7\": rpc error: code = NotFound desc = could not find container \"4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7\": container with ID starting with 4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7 not found: ID does not exist" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.190925 4856 scope.go:117] "RemoveContainer" containerID="b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.191602 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde"} err="failed to get container status \"b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde\": rpc error: code = NotFound desc = could not find container \"b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde\": container with ID starting with b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde not found: ID does not exist" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.191624 4856 scope.go:117] "RemoveContainer" containerID="7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.193115 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3"} err="failed to get container status \"7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3\": rpc error: code = NotFound desc = could not find container \"7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3\": container with ID starting with 7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3 not found: ID does not exist" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.193263 4856 scope.go:117] "RemoveContainer" containerID="e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.194231 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1"} err="failed to get container status \"e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1\": rpc error: code = NotFound desc = could not find container \"e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1\": container with ID starting with e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1 not found: ID does not exist" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.194257 4856 scope.go:117] "RemoveContainer" containerID="83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.195225 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc"} err="failed to get container status \"83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc\": rpc error: code = NotFound desc = could not find container \"83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc\": container with ID starting with 83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc not found: ID does not exist" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.195257 4856 scope.go:117] "RemoveContainer" containerID="25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.195709 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b"} err="failed to get container status \"25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b\": rpc error: code = NotFound desc = could not find container \"25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b\": container with ID starting with 25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b not found: ID does not exist" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.195732 4856 scope.go:117] "RemoveContainer" containerID="d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.196201 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8"} err="failed to get container status \"d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\": rpc error: code = NotFound desc = could not find container \"d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\": container with ID starting with d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8 not found: ID does not exist" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.196229 4856 scope.go:117] "RemoveContainer" containerID="984517fa62d3e73cf4b33b7c4a101f8221c940f66938120373b7d41cac9c5e5c" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.199011 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"984517fa62d3e73cf4b33b7c4a101f8221c940f66938120373b7d41cac9c5e5c"} err="failed to get container status \"984517fa62d3e73cf4b33b7c4a101f8221c940f66938120373b7d41cac9c5e5c\": rpc error: code = NotFound desc = could not find container \"984517fa62d3e73cf4b33b7c4a101f8221c940f66938120373b7d41cac9c5e5c\": container with ID starting with 984517fa62d3e73cf4b33b7c4a101f8221c940f66938120373b7d41cac9c5e5c not found: ID does not exist" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.199193 4856 scope.go:117] "RemoveContainer" containerID="203d756903498bed1c57a3e87a95f4b24f808514567a99505ba2f3cfec468cb6" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.199819 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"203d756903498bed1c57a3e87a95f4b24f808514567a99505ba2f3cfec468cb6"} err="failed to get container status \"203d756903498bed1c57a3e87a95f4b24f808514567a99505ba2f3cfec468cb6\": rpc error: code = NotFound desc = could not find container \"203d756903498bed1c57a3e87a95f4b24f808514567a99505ba2f3cfec468cb6\": container with ID starting with 203d756903498bed1c57a3e87a95f4b24f808514567a99505ba2f3cfec468cb6 not found: ID does not exist" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.199915 4856 scope.go:117] "RemoveContainer" containerID="11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.200306 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a"} err="failed to get container status \"11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a\": rpc error: code = NotFound desc = could not find container \"11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a\": container with ID starting with 11205cc9ebaa35eb58159e387e540fdb6fa8a75b628b5f1b1e79e640665ced4a not found: ID does not exist" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.200331 4856 scope.go:117] "RemoveContainer" containerID="4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.200594 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7"} err="failed to get container status \"4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7\": rpc error: code = NotFound desc = could not find container \"4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7\": container with ID starting with 4a1a575c02f04d857f279139bf41de49cbf2ad326a9e63be6757b7fe72dd26d7 not found: ID does not exist" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.200702 4856 scope.go:117] "RemoveContainer" containerID="b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.201098 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde"} err="failed to get container status \"b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde\": rpc error: code = NotFound desc = could not find container \"b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde\": container with ID starting with b1509fb370f6f5002154d55db3aa12e20b4aaef2552faccc0bc6e22378a28fde not found: ID does not exist" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.201121 4856 scope.go:117] "RemoveContainer" containerID="7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.201486 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3"} err="failed to get container status \"7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3\": rpc error: code = NotFound desc = could not find container \"7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3\": container with ID starting with 7f3ce6d59efe2830eff50e1ff6deb9464d70926fa6b937c4fa325b2f6c82cca3 not found: ID does not exist" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.201598 4856 scope.go:117] "RemoveContainer" containerID="e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.201931 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1"} err="failed to get container status \"e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1\": rpc error: code = NotFound desc = could not find container \"e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1\": container with ID starting with e9f7a132d7600cfb27a4fa2a46f8d3469faa1f4f3792f99884ca456fa4aa71b1 not found: ID does not exist" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.202027 4856 scope.go:117] "RemoveContainer" containerID="83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.202506 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc"} err="failed to get container status \"83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc\": rpc error: code = NotFound desc = could not find container \"83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc\": container with ID starting with 83a08d4efc6e12956b4420eeec79e50e426f29b90bcf50edadceddc1718d88fc not found: ID does not exist" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.202647 4856 scope.go:117] "RemoveContainer" containerID="25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.203110 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b"} err="failed to get container status \"25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b\": rpc error: code = NotFound desc = could not find container \"25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b\": container with ID starting with 25056a87b6686fa3172581a9c5a889f286abe72ac2acfe81463e072deb9e850b not found: ID does not exist" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.203200 4856 scope.go:117] "RemoveContainer" containerID="d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.203566 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8"} err="failed to get container status \"d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\": rpc error: code = NotFound desc = could not find container \"d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8\": container with ID starting with d67317281647e2a2cf1b88f162a6c3cc224c243ff7a13f9706f416ca0e45dee8 not found: ID does not exist" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.203598 4856 scope.go:117] "RemoveContainer" containerID="984517fa62d3e73cf4b33b7c4a101f8221c940f66938120373b7d41cac9c5e5c" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.203856 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"984517fa62d3e73cf4b33b7c4a101f8221c940f66938120373b7d41cac9c5e5c"} err="failed to get container status \"984517fa62d3e73cf4b33b7c4a101f8221c940f66938120373b7d41cac9c5e5c\": rpc error: code = NotFound desc = could not find container \"984517fa62d3e73cf4b33b7c4a101f8221c940f66938120373b7d41cac9c5e5c\": container with ID starting with 984517fa62d3e73cf4b33b7c4a101f8221c940f66938120373b7d41cac9c5e5c not found: ID does not exist" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.403469 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab5b6f50-172b-4535-a0f9-5d103bcab4e7" path="/var/lib/kubelet/pods/ab5b6f50-172b-4535-a0f9-5d103bcab4e7/volumes" Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.990968 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" event={"ID":"d70f72f7-ff5e-4906-8622-9cddfe769d55","Type":"ContainerStarted","Data":"91ef65730af9e2bdd7621585a901cb68f84170da5efdf43be4173574d0ab23e3"} Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.991316 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" event={"ID":"d70f72f7-ff5e-4906-8622-9cddfe769d55","Type":"ContainerStarted","Data":"0c74ebe4a236a613e6e7856badc6e52a30604cfa5668c23fb59ca1b7570a3c8a"} Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.991333 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" event={"ID":"d70f72f7-ff5e-4906-8622-9cddfe769d55","Type":"ContainerStarted","Data":"fa2f5e2ecee3ae541cd01ea593fb352363c772d981279fd1b557036b071363a0"} Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.991342 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" event={"ID":"d70f72f7-ff5e-4906-8622-9cddfe769d55","Type":"ContainerStarted","Data":"29faba44ad25215ae335927fbfb28db292d80b8727d2b69e64f52a531eba697a"} Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.991351 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" event={"ID":"d70f72f7-ff5e-4906-8622-9cddfe769d55","Type":"ContainerStarted","Data":"37fc5e527ffa138328aa0bc335a3685420067c147999a07f5198c6168d40b84c"} Jan 26 17:09:19 crc kubenswrapper[4856]: I0126 17:09:19.991360 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" event={"ID":"d70f72f7-ff5e-4906-8622-9cddfe769d55","Type":"ContainerStarted","Data":"3e07b4c3b62de15e63aaa9d185f89327d5da32a85c68e5c16a0217ac747b5fcc"} Jan 26 17:09:22 crc kubenswrapper[4856]: I0126 17:09:22.008599 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" event={"ID":"d70f72f7-ff5e-4906-8622-9cddfe769d55","Type":"ContainerStarted","Data":"3c8af924f05f38bec8960462460d29ef0d36c059c00c95719be5ab41276cb331"} Jan 26 17:09:25 crc kubenswrapper[4856]: I0126 17:09:25.030291 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" event={"ID":"d70f72f7-ff5e-4906-8622-9cddfe769d55","Type":"ContainerStarted","Data":"b34423efb2d1e07043ed5f4c4bb31cedabbbaf0d8733a1651fbab70cc95a3c6c"} Jan 26 17:09:25 crc kubenswrapper[4856]: I0126 17:09:25.030827 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:25 crc kubenswrapper[4856]: I0126 17:09:25.030840 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:25 crc kubenswrapper[4856]: I0126 17:09:25.030849 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:25 crc kubenswrapper[4856]: I0126 17:09:25.054355 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:25 crc kubenswrapper[4856]: I0126 17:09:25.059277 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" podStartSLOduration=7.059249732 podStartE2EDuration="7.059249732s" podCreationTimestamp="2026-01-26 17:09:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:09:25.057688437 +0000 UTC m=+661.010942428" watchObservedRunningTime="2026-01-26 17:09:25.059249732 +0000 UTC m=+661.012503723" Jan 26 17:09:25 crc kubenswrapper[4856]: I0126 17:09:25.060962 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:09:25 crc kubenswrapper[4856]: I0126 17:09:25.827083 4856 scope.go:117] "RemoveContainer" containerID="67c41d7af13d33af9423d069e86b531ff9d226b1435b62347517f490f3904943" Jan 26 17:09:25 crc kubenswrapper[4856]: I0126 17:09:25.848408 4856 scope.go:117] "RemoveContainer" containerID="d684603f69b61a1ce87ec7d1d3ef00e518372571ee64ede6a51ce75afd2227ca" Jan 26 17:09:25 crc kubenswrapper[4856]: I0126 17:09:25.869358 4856 scope.go:117] "RemoveContainer" containerID="afeb20035224feeab28a92ac77b43a24e653e49c56a25590a9861019a2b7a8ff" Jan 26 17:09:26 crc kubenswrapper[4856]: I0126 17:09:26.939201 4856 patch_prober.go:28] interesting pod/machine-config-daemon-xm9cq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:09:26 crc kubenswrapper[4856]: I0126 17:09:26.941228 4856 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:09:26 crc kubenswrapper[4856]: I0126 17:09:26.941466 4856 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" Jan 26 17:09:26 crc kubenswrapper[4856]: I0126 17:09:26.942616 4856 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fe42c0299ac9f35a2260caaf7226f7e2161da013442117dab0d25a7c69c46115"} pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 26 17:09:26 crc kubenswrapper[4856]: I0126 17:09:26.942920 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" containerName="machine-config-daemon" containerID="cri-o://fe42c0299ac9f35a2260caaf7226f7e2161da013442117dab0d25a7c69c46115" gracePeriod=600 Jan 26 17:09:27 crc kubenswrapper[4856]: I0126 17:09:27.044610 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-rq622_7a742e7b-c420-46e3-9e96-e9c744af6124/kube-multus/2.log" Jan 26 17:09:28 crc kubenswrapper[4856]: I0126 17:09:28.054728 4856 generic.go:334] "Generic (PLEG): container finished" podID="63c75ede-5170-4db0-811b-5217ef8d72b3" containerID="fe42c0299ac9f35a2260caaf7226f7e2161da013442117dab0d25a7c69c46115" exitCode=0 Jan 26 17:09:28 crc kubenswrapper[4856]: I0126 17:09:28.054848 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" event={"ID":"63c75ede-5170-4db0-811b-5217ef8d72b3","Type":"ContainerDied","Data":"fe42c0299ac9f35a2260caaf7226f7e2161da013442117dab0d25a7c69c46115"} Jan 26 17:09:28 crc kubenswrapper[4856]: I0126 17:09:28.055067 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" event={"ID":"63c75ede-5170-4db0-811b-5217ef8d72b3","Type":"ContainerStarted","Data":"bb3fb578d0ea2b4eb264b402043faa4d1923f5d38749a2ee2c65b084c2e291bd"} Jan 26 17:09:28 crc kubenswrapper[4856]: I0126 17:09:28.055117 4856 scope.go:117] "RemoveContainer" containerID="9758bfdfd1807e791935ac7ec93246863e5867351e35d27ffaff68ae79110e9c" Jan 26 17:09:34 crc kubenswrapper[4856]: I0126 17:09:34.396001 4856 scope.go:117] "RemoveContainer" containerID="ddec0dbea657c6160cfdfd78886d5ae335dab8b667b0e0e3813dffa86a2ae2dc" Jan 26 17:09:34 crc kubenswrapper[4856]: E0126 17:09:34.396945 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-rq622_openshift-multus(7a742e7b-c420-46e3-9e96-e9c744af6124)\"" pod="openshift-multus/multus-rq622" podUID="7a742e7b-c420-46e3-9e96-e9c744af6124" Jan 26 17:09:46 crc kubenswrapper[4856]: I0126 17:09:46.395383 4856 scope.go:117] "RemoveContainer" containerID="ddec0dbea657c6160cfdfd78886d5ae335dab8b667b0e0e3813dffa86a2ae2dc" Jan 26 17:09:47 crc kubenswrapper[4856]: I0126 17:09:47.189977 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-rq622_7a742e7b-c420-46e3-9e96-e9c744af6124/kube-multus/2.log" Jan 26 17:09:47 crc kubenswrapper[4856]: I0126 17:09:47.190343 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-rq622" event={"ID":"7a742e7b-c420-46e3-9e96-e9c744af6124","Type":"ContainerStarted","Data":"443165ce1d5496709cff016aaa51725cac9a85718dc182fc9666e5c69f45c262"} Jan 26 17:09:48 crc kubenswrapper[4856]: I0126 17:09:48.638361 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6t25x" Jan 26 17:10:39 crc kubenswrapper[4856]: I0126 17:10:39.699767 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-97mff"] Jan 26 17:10:39 crc kubenswrapper[4856]: I0126 17:10:39.702516 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-97mff" podUID="886857c0-659b-4904-b75a-c55c3f712747" containerName="registry-server" containerID="cri-o://86283045d7d1049d9d8358f985c6aac8c275ef1f0b7a9715b13fd30bd1c328e7" gracePeriod=30 Jan 26 17:10:40 crc kubenswrapper[4856]: I0126 17:10:40.019857 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-97mff" Jan 26 17:10:40 crc kubenswrapper[4856]: I0126 17:10:40.209385 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9487\" (UniqueName: \"kubernetes.io/projected/886857c0-659b-4904-b75a-c55c3f712747-kube-api-access-q9487\") pod \"886857c0-659b-4904-b75a-c55c3f712747\" (UID: \"886857c0-659b-4904-b75a-c55c3f712747\") " Jan 26 17:10:40 crc kubenswrapper[4856]: I0126 17:10:40.209503 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/886857c0-659b-4904-b75a-c55c3f712747-utilities\") pod \"886857c0-659b-4904-b75a-c55c3f712747\" (UID: \"886857c0-659b-4904-b75a-c55c3f712747\") " Jan 26 17:10:40 crc kubenswrapper[4856]: I0126 17:10:40.209629 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/886857c0-659b-4904-b75a-c55c3f712747-catalog-content\") pod \"886857c0-659b-4904-b75a-c55c3f712747\" (UID: \"886857c0-659b-4904-b75a-c55c3f712747\") " Jan 26 17:10:40 crc kubenswrapper[4856]: I0126 17:10:40.210884 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/886857c0-659b-4904-b75a-c55c3f712747-utilities" (OuterVolumeSpecName: "utilities") pod "886857c0-659b-4904-b75a-c55c3f712747" (UID: "886857c0-659b-4904-b75a-c55c3f712747"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:10:40 crc kubenswrapper[4856]: I0126 17:10:40.217815 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/886857c0-659b-4904-b75a-c55c3f712747-kube-api-access-q9487" (OuterVolumeSpecName: "kube-api-access-q9487") pod "886857c0-659b-4904-b75a-c55c3f712747" (UID: "886857c0-659b-4904-b75a-c55c3f712747"). InnerVolumeSpecName "kube-api-access-q9487". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:10:40 crc kubenswrapper[4856]: I0126 17:10:40.234950 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/886857c0-659b-4904-b75a-c55c3f712747-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "886857c0-659b-4904-b75a-c55c3f712747" (UID: "886857c0-659b-4904-b75a-c55c3f712747"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:10:40 crc kubenswrapper[4856]: I0126 17:10:40.311490 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9487\" (UniqueName: \"kubernetes.io/projected/886857c0-659b-4904-b75a-c55c3f712747-kube-api-access-q9487\") on node \"crc\" DevicePath \"\"" Jan 26 17:10:40 crc kubenswrapper[4856]: I0126 17:10:40.311574 4856 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/886857c0-659b-4904-b75a-c55c3f712747-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 17:10:40 crc kubenswrapper[4856]: I0126 17:10:40.311588 4856 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/886857c0-659b-4904-b75a-c55c3f712747-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 17:10:40 crc kubenswrapper[4856]: I0126 17:10:40.554503 4856 generic.go:334] "Generic (PLEG): container finished" podID="886857c0-659b-4904-b75a-c55c3f712747" containerID="86283045d7d1049d9d8358f985c6aac8c275ef1f0b7a9715b13fd30bd1c328e7" exitCode=0 Jan 26 17:10:40 crc kubenswrapper[4856]: I0126 17:10:40.554890 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-97mff" event={"ID":"886857c0-659b-4904-b75a-c55c3f712747","Type":"ContainerDied","Data":"86283045d7d1049d9d8358f985c6aac8c275ef1f0b7a9715b13fd30bd1c328e7"} Jan 26 17:10:40 crc kubenswrapper[4856]: I0126 17:10:40.555091 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-97mff" event={"ID":"886857c0-659b-4904-b75a-c55c3f712747","Type":"ContainerDied","Data":"2063535f537d4fe37e3e34708f04c20619c6cc50b85697e69a9333b26c91a793"} Jan 26 17:10:40 crc kubenswrapper[4856]: I0126 17:10:40.555360 4856 scope.go:117] "RemoveContainer" containerID="86283045d7d1049d9d8358f985c6aac8c275ef1f0b7a9715b13fd30bd1c328e7" Jan 26 17:10:40 crc kubenswrapper[4856]: I0126 17:10:40.555804 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-97mff" Jan 26 17:10:40 crc kubenswrapper[4856]: I0126 17:10:40.578098 4856 scope.go:117] "RemoveContainer" containerID="a75ef75367730507a8b7594226c5e9d4e14716073f574dda81c029b084dafd94" Jan 26 17:10:40 crc kubenswrapper[4856]: I0126 17:10:40.600375 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-97mff"] Jan 26 17:10:40 crc kubenswrapper[4856]: I0126 17:10:40.602740 4856 scope.go:117] "RemoveContainer" containerID="8cce484e79d411777eb43ce1a40864e7613f816cb566efdd41677d117f9c3633" Jan 26 17:10:40 crc kubenswrapper[4856]: I0126 17:10:40.604927 4856 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-97mff"] Jan 26 17:10:40 crc kubenswrapper[4856]: I0126 17:10:40.617053 4856 scope.go:117] "RemoveContainer" containerID="86283045d7d1049d9d8358f985c6aac8c275ef1f0b7a9715b13fd30bd1c328e7" Jan 26 17:10:40 crc kubenswrapper[4856]: E0126 17:10:40.617617 4856 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86283045d7d1049d9d8358f985c6aac8c275ef1f0b7a9715b13fd30bd1c328e7\": container with ID starting with 86283045d7d1049d9d8358f985c6aac8c275ef1f0b7a9715b13fd30bd1c328e7 not found: ID does not exist" containerID="86283045d7d1049d9d8358f985c6aac8c275ef1f0b7a9715b13fd30bd1c328e7" Jan 26 17:10:40 crc kubenswrapper[4856]: I0126 17:10:40.617767 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86283045d7d1049d9d8358f985c6aac8c275ef1f0b7a9715b13fd30bd1c328e7"} err="failed to get container status \"86283045d7d1049d9d8358f985c6aac8c275ef1f0b7a9715b13fd30bd1c328e7\": rpc error: code = NotFound desc = could not find container \"86283045d7d1049d9d8358f985c6aac8c275ef1f0b7a9715b13fd30bd1c328e7\": container with ID starting with 86283045d7d1049d9d8358f985c6aac8c275ef1f0b7a9715b13fd30bd1c328e7 not found: ID does not exist" Jan 26 17:10:40 crc kubenswrapper[4856]: I0126 17:10:40.617859 4856 scope.go:117] "RemoveContainer" containerID="a75ef75367730507a8b7594226c5e9d4e14716073f574dda81c029b084dafd94" Jan 26 17:10:40 crc kubenswrapper[4856]: E0126 17:10:40.618331 4856 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a75ef75367730507a8b7594226c5e9d4e14716073f574dda81c029b084dafd94\": container with ID starting with a75ef75367730507a8b7594226c5e9d4e14716073f574dda81c029b084dafd94 not found: ID does not exist" containerID="a75ef75367730507a8b7594226c5e9d4e14716073f574dda81c029b084dafd94" Jan 26 17:10:40 crc kubenswrapper[4856]: I0126 17:10:40.618422 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a75ef75367730507a8b7594226c5e9d4e14716073f574dda81c029b084dafd94"} err="failed to get container status \"a75ef75367730507a8b7594226c5e9d4e14716073f574dda81c029b084dafd94\": rpc error: code = NotFound desc = could not find container \"a75ef75367730507a8b7594226c5e9d4e14716073f574dda81c029b084dafd94\": container with ID starting with a75ef75367730507a8b7594226c5e9d4e14716073f574dda81c029b084dafd94 not found: ID does not exist" Jan 26 17:10:40 crc kubenswrapper[4856]: I0126 17:10:40.618484 4856 scope.go:117] "RemoveContainer" containerID="8cce484e79d411777eb43ce1a40864e7613f816cb566efdd41677d117f9c3633" Jan 26 17:10:40 crc kubenswrapper[4856]: E0126 17:10:40.619030 4856 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8cce484e79d411777eb43ce1a40864e7613f816cb566efdd41677d117f9c3633\": container with ID starting with 8cce484e79d411777eb43ce1a40864e7613f816cb566efdd41677d117f9c3633 not found: ID does not exist" containerID="8cce484e79d411777eb43ce1a40864e7613f816cb566efdd41677d117f9c3633" Jan 26 17:10:40 crc kubenswrapper[4856]: I0126 17:10:40.619086 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cce484e79d411777eb43ce1a40864e7613f816cb566efdd41677d117f9c3633"} err="failed to get container status \"8cce484e79d411777eb43ce1a40864e7613f816cb566efdd41677d117f9c3633\": rpc error: code = NotFound desc = could not find container \"8cce484e79d411777eb43ce1a40864e7613f816cb566efdd41677d117f9c3633\": container with ID starting with 8cce484e79d411777eb43ce1a40864e7613f816cb566efdd41677d117f9c3633 not found: ID does not exist" Jan 26 17:10:41 crc kubenswrapper[4856]: I0126 17:10:41.406272 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="886857c0-659b-4904-b75a-c55c3f712747" path="/var/lib/kubelet/pods/886857c0-659b-4904-b75a-c55c3f712747/volumes" Jan 26 17:10:43 crc kubenswrapper[4856]: I0126 17:10:43.922960 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08924r9"] Jan 26 17:10:43 crc kubenswrapper[4856]: E0126 17:10:43.923704 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="886857c0-659b-4904-b75a-c55c3f712747" containerName="extract-content" Jan 26 17:10:43 crc kubenswrapper[4856]: I0126 17:10:43.923728 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="886857c0-659b-4904-b75a-c55c3f712747" containerName="extract-content" Jan 26 17:10:43 crc kubenswrapper[4856]: E0126 17:10:43.923746 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="886857c0-659b-4904-b75a-c55c3f712747" containerName="extract-utilities" Jan 26 17:10:43 crc kubenswrapper[4856]: I0126 17:10:43.923760 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="886857c0-659b-4904-b75a-c55c3f712747" containerName="extract-utilities" Jan 26 17:10:43 crc kubenswrapper[4856]: E0126 17:10:43.923770 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="886857c0-659b-4904-b75a-c55c3f712747" containerName="registry-server" Jan 26 17:10:43 crc kubenswrapper[4856]: I0126 17:10:43.923777 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="886857c0-659b-4904-b75a-c55c3f712747" containerName="registry-server" Jan 26 17:10:43 crc kubenswrapper[4856]: I0126 17:10:43.923980 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="886857c0-659b-4904-b75a-c55c3f712747" containerName="registry-server" Jan 26 17:10:43 crc kubenswrapper[4856]: I0126 17:10:43.925109 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08924r9" Jan 26 17:10:43 crc kubenswrapper[4856]: I0126 17:10:43.928247 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Jan 26 17:10:43 crc kubenswrapper[4856]: I0126 17:10:43.937904 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08924r9"] Jan 26 17:10:44 crc kubenswrapper[4856]: I0126 17:10:44.001327 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bfqg\" (UniqueName: \"kubernetes.io/projected/64c65d72-3459-4893-a33a-9033e12f188a-kube-api-access-8bfqg\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08924r9\" (UID: \"64c65d72-3459-4893-a33a-9033e12f188a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08924r9" Jan 26 17:10:44 crc kubenswrapper[4856]: I0126 17:10:44.001444 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/64c65d72-3459-4893-a33a-9033e12f188a-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08924r9\" (UID: \"64c65d72-3459-4893-a33a-9033e12f188a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08924r9" Jan 26 17:10:44 crc kubenswrapper[4856]: I0126 17:10:44.001494 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/64c65d72-3459-4893-a33a-9033e12f188a-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08924r9\" (UID: \"64c65d72-3459-4893-a33a-9033e12f188a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08924r9" Jan 26 17:10:44 crc kubenswrapper[4856]: I0126 17:10:44.102601 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bfqg\" (UniqueName: \"kubernetes.io/projected/64c65d72-3459-4893-a33a-9033e12f188a-kube-api-access-8bfqg\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08924r9\" (UID: \"64c65d72-3459-4893-a33a-9033e12f188a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08924r9" Jan 26 17:10:44 crc kubenswrapper[4856]: I0126 17:10:44.103208 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/64c65d72-3459-4893-a33a-9033e12f188a-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08924r9\" (UID: \"64c65d72-3459-4893-a33a-9033e12f188a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08924r9" Jan 26 17:10:44 crc kubenswrapper[4856]: I0126 17:10:44.103368 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/64c65d72-3459-4893-a33a-9033e12f188a-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08924r9\" (UID: \"64c65d72-3459-4893-a33a-9033e12f188a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08924r9" Jan 26 17:10:44 crc kubenswrapper[4856]: I0126 17:10:44.103786 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/64c65d72-3459-4893-a33a-9033e12f188a-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08924r9\" (UID: \"64c65d72-3459-4893-a33a-9033e12f188a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08924r9" Jan 26 17:10:44 crc kubenswrapper[4856]: I0126 17:10:44.103890 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/64c65d72-3459-4893-a33a-9033e12f188a-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08924r9\" (UID: \"64c65d72-3459-4893-a33a-9033e12f188a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08924r9" Jan 26 17:10:44 crc kubenswrapper[4856]: I0126 17:10:44.123131 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bfqg\" (UniqueName: \"kubernetes.io/projected/64c65d72-3459-4893-a33a-9033e12f188a-kube-api-access-8bfqg\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08924r9\" (UID: \"64c65d72-3459-4893-a33a-9033e12f188a\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08924r9" Jan 26 17:10:44 crc kubenswrapper[4856]: I0126 17:10:44.241776 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08924r9" Jan 26 17:10:44 crc kubenswrapper[4856]: I0126 17:10:44.417347 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08924r9"] Jan 26 17:10:44 crc kubenswrapper[4856]: I0126 17:10:44.602961 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08924r9" event={"ID":"64c65d72-3459-4893-a33a-9033e12f188a","Type":"ContainerStarted","Data":"89890338723d888ef2be71ab9569ddc46c833b10c10f6fb75ebe4a541095d7fa"} Jan 26 17:10:44 crc kubenswrapper[4856]: I0126 17:10:44.603012 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08924r9" event={"ID":"64c65d72-3459-4893-a33a-9033e12f188a","Type":"ContainerStarted","Data":"3f751138bd790ab355d7a769821c2ba507a3c1d559249b5a974c3d17ab565a17"} Jan 26 17:10:45 crc kubenswrapper[4856]: I0126 17:10:45.610376 4856 generic.go:334] "Generic (PLEG): container finished" podID="64c65d72-3459-4893-a33a-9033e12f188a" containerID="89890338723d888ef2be71ab9569ddc46c833b10c10f6fb75ebe4a541095d7fa" exitCode=0 Jan 26 17:10:45 crc kubenswrapper[4856]: I0126 17:10:45.610766 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08924r9" event={"ID":"64c65d72-3459-4893-a33a-9033e12f188a","Type":"ContainerDied","Data":"89890338723d888ef2be71ab9569ddc46c833b10c10f6fb75ebe4a541095d7fa"} Jan 26 17:10:45 crc kubenswrapper[4856]: I0126 17:10:45.615103 4856 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 26 17:10:49 crc kubenswrapper[4856]: I0126 17:10:49.634320 4856 generic.go:334] "Generic (PLEG): container finished" podID="64c65d72-3459-4893-a33a-9033e12f188a" containerID="78251e21df888156bd730c79137b6f0a500a76e4efd92c48da4997db63024cb0" exitCode=0 Jan 26 17:10:49 crc kubenswrapper[4856]: I0126 17:10:49.634441 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08924r9" event={"ID":"64c65d72-3459-4893-a33a-9033e12f188a","Type":"ContainerDied","Data":"78251e21df888156bd730c79137b6f0a500a76e4efd92c48da4997db63024cb0"} Jan 26 17:10:50 crc kubenswrapper[4856]: I0126 17:10:50.642907 4856 generic.go:334] "Generic (PLEG): container finished" podID="64c65d72-3459-4893-a33a-9033e12f188a" containerID="1881326ff739ce94ce545803f5d57dc24b0f53ff04a2bbb6f44ae945527f62ea" exitCode=0 Jan 26 17:10:50 crc kubenswrapper[4856]: I0126 17:10:50.642961 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08924r9" event={"ID":"64c65d72-3459-4893-a33a-9033e12f188a","Type":"ContainerDied","Data":"1881326ff739ce94ce545803f5d57dc24b0f53ff04a2bbb6f44ae945527f62ea"} Jan 26 17:10:51 crc kubenswrapper[4856]: I0126 17:10:51.916142 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08924r9" Jan 26 17:10:51 crc kubenswrapper[4856]: I0126 17:10:51.931228 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/64c65d72-3459-4893-a33a-9033e12f188a-util\") pod \"64c65d72-3459-4893-a33a-9033e12f188a\" (UID: \"64c65d72-3459-4893-a33a-9033e12f188a\") " Jan 26 17:10:51 crc kubenswrapper[4856]: I0126 17:10:51.931386 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8bfqg\" (UniqueName: \"kubernetes.io/projected/64c65d72-3459-4893-a33a-9033e12f188a-kube-api-access-8bfqg\") pod \"64c65d72-3459-4893-a33a-9033e12f188a\" (UID: \"64c65d72-3459-4893-a33a-9033e12f188a\") " Jan 26 17:10:51 crc kubenswrapper[4856]: I0126 17:10:51.931449 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/64c65d72-3459-4893-a33a-9033e12f188a-bundle\") pod \"64c65d72-3459-4893-a33a-9033e12f188a\" (UID: \"64c65d72-3459-4893-a33a-9033e12f188a\") " Jan 26 17:10:51 crc kubenswrapper[4856]: I0126 17:10:51.939281 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64c65d72-3459-4893-a33a-9033e12f188a-bundle" (OuterVolumeSpecName: "bundle") pod "64c65d72-3459-4893-a33a-9033e12f188a" (UID: "64c65d72-3459-4893-a33a-9033e12f188a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:10:51 crc kubenswrapper[4856]: I0126 17:10:51.946250 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64c65d72-3459-4893-a33a-9033e12f188a-util" (OuterVolumeSpecName: "util") pod "64c65d72-3459-4893-a33a-9033e12f188a" (UID: "64c65d72-3459-4893-a33a-9033e12f188a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:10:51 crc kubenswrapper[4856]: I0126 17:10:51.946711 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64c65d72-3459-4893-a33a-9033e12f188a-kube-api-access-8bfqg" (OuterVolumeSpecName: "kube-api-access-8bfqg") pod "64c65d72-3459-4893-a33a-9033e12f188a" (UID: "64c65d72-3459-4893-a33a-9033e12f188a"). InnerVolumeSpecName "kube-api-access-8bfqg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:10:52 crc kubenswrapper[4856]: I0126 17:10:52.033350 4856 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/64c65d72-3459-4893-a33a-9033e12f188a-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:10:52 crc kubenswrapper[4856]: I0126 17:10:52.033390 4856 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/64c65d72-3459-4893-a33a-9033e12f188a-util\") on node \"crc\" DevicePath \"\"" Jan 26 17:10:52 crc kubenswrapper[4856]: I0126 17:10:52.033400 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8bfqg\" (UniqueName: \"kubernetes.io/projected/64c65d72-3459-4893-a33a-9033e12f188a-kube-api-access-8bfqg\") on node \"crc\" DevicePath \"\"" Jan 26 17:10:52 crc kubenswrapper[4856]: I0126 17:10:52.673786 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5e25mvq"] Jan 26 17:10:52 crc kubenswrapper[4856]: E0126 17:10:52.674113 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64c65d72-3459-4893-a33a-9033e12f188a" containerName="util" Jan 26 17:10:52 crc kubenswrapper[4856]: I0126 17:10:52.674136 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="64c65d72-3459-4893-a33a-9033e12f188a" containerName="util" Jan 26 17:10:52 crc kubenswrapper[4856]: E0126 17:10:52.674150 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64c65d72-3459-4893-a33a-9033e12f188a" containerName="extract" Jan 26 17:10:52 crc kubenswrapper[4856]: I0126 17:10:52.674160 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="64c65d72-3459-4893-a33a-9033e12f188a" containerName="extract" Jan 26 17:10:52 crc kubenswrapper[4856]: E0126 17:10:52.674177 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64c65d72-3459-4893-a33a-9033e12f188a" containerName="pull" Jan 26 17:10:52 crc kubenswrapper[4856]: I0126 17:10:52.674185 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="64c65d72-3459-4893-a33a-9033e12f188a" containerName="pull" Jan 26 17:10:52 crc kubenswrapper[4856]: I0126 17:10:52.674314 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="64c65d72-3459-4893-a33a-9033e12f188a" containerName="extract" Jan 26 17:10:52 crc kubenswrapper[4856]: I0126 17:10:52.675258 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5e25mvq" Jan 26 17:10:52 crc kubenswrapper[4856]: I0126 17:10:52.684891 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5e25mvq"] Jan 26 17:10:52 crc kubenswrapper[4856]: I0126 17:10:52.707561 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08924r9" event={"ID":"64c65d72-3459-4893-a33a-9033e12f188a","Type":"ContainerDied","Data":"3f751138bd790ab355d7a769821c2ba507a3c1d559249b5a974c3d17ab565a17"} Jan 26 17:10:52 crc kubenswrapper[4856]: I0126 17:10:52.707604 4856 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f751138bd790ab355d7a769821c2ba507a3c1d559249b5a974c3d17ab565a17" Jan 26 17:10:52 crc kubenswrapper[4856]: I0126 17:10:52.707641 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08924r9" Jan 26 17:10:52 crc kubenswrapper[4856]: I0126 17:10:52.742816 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a-util\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5e25mvq\" (UID: \"26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5e25mvq" Jan 26 17:10:52 crc kubenswrapper[4856]: I0126 17:10:52.743007 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2wfr\" (UniqueName: \"kubernetes.io/projected/26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a-kube-api-access-j2wfr\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5e25mvq\" (UID: \"26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5e25mvq" Jan 26 17:10:52 crc kubenswrapper[4856]: I0126 17:10:52.743322 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a-bundle\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5e25mvq\" (UID: \"26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5e25mvq" Jan 26 17:10:52 crc kubenswrapper[4856]: I0126 17:10:52.845034 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a-util\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5e25mvq\" (UID: \"26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5e25mvq" Jan 26 17:10:52 crc kubenswrapper[4856]: I0126 17:10:52.845110 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2wfr\" (UniqueName: \"kubernetes.io/projected/26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a-kube-api-access-j2wfr\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5e25mvq\" (UID: \"26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5e25mvq" Jan 26 17:10:52 crc kubenswrapper[4856]: I0126 17:10:52.845141 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a-bundle\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5e25mvq\" (UID: \"26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5e25mvq" Jan 26 17:10:52 crc kubenswrapper[4856]: I0126 17:10:52.846158 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a-bundle\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5e25mvq\" (UID: \"26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5e25mvq" Jan 26 17:10:52 crc kubenswrapper[4856]: I0126 17:10:52.846152 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a-util\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5e25mvq\" (UID: \"26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5e25mvq" Jan 26 17:10:52 crc kubenswrapper[4856]: I0126 17:10:52.870678 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2wfr\" (UniqueName: \"kubernetes.io/projected/26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a-kube-api-access-j2wfr\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5e25mvq\" (UID: \"26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5e25mvq" Jan 26 17:10:52 crc kubenswrapper[4856]: I0126 17:10:52.990753 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5e25mvq" Jan 26 17:10:53 crc kubenswrapper[4856]: I0126 17:10:53.168064 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5e25mvq"] Jan 26 17:10:53 crc kubenswrapper[4856]: I0126 17:10:53.671130 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fxrgwg"] Jan 26 17:10:53 crc kubenswrapper[4856]: I0126 17:10:53.673034 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fxrgwg" Jan 26 17:10:53 crc kubenswrapper[4856]: I0126 17:10:53.684729 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fxrgwg"] Jan 26 17:10:53 crc kubenswrapper[4856]: I0126 17:10:53.713397 4856 generic.go:334] "Generic (PLEG): container finished" podID="26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a" containerID="c757a6d599d238252cf69718c03a250c4bd64e682f5b5456a36c0a4aa37edbc8" exitCode=0 Jan 26 17:10:53 crc kubenswrapper[4856]: I0126 17:10:53.713439 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5e25mvq" event={"ID":"26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a","Type":"ContainerDied","Data":"c757a6d599d238252cf69718c03a250c4bd64e682f5b5456a36c0a4aa37edbc8"} Jan 26 17:10:53 crc kubenswrapper[4856]: I0126 17:10:53.713466 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5e25mvq" event={"ID":"26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a","Type":"ContainerStarted","Data":"13128ab773dafbdc20c6ca5345fca45949f139a13824371598c28655f37fe918"} Jan 26 17:10:53 crc kubenswrapper[4856]: I0126 17:10:53.756143 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7105e655-ab8e-4fc0-b205-0bafaa6d7d91-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fxrgwg\" (UID: \"7105e655-ab8e-4fc0-b205-0bafaa6d7d91\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fxrgwg" Jan 26 17:10:53 crc kubenswrapper[4856]: I0126 17:10:53.756334 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tb25l\" (UniqueName: \"kubernetes.io/projected/7105e655-ab8e-4fc0-b205-0bafaa6d7d91-kube-api-access-tb25l\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fxrgwg\" (UID: \"7105e655-ab8e-4fc0-b205-0bafaa6d7d91\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fxrgwg" Jan 26 17:10:53 crc kubenswrapper[4856]: I0126 17:10:53.756398 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7105e655-ab8e-4fc0-b205-0bafaa6d7d91-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fxrgwg\" (UID: \"7105e655-ab8e-4fc0-b205-0bafaa6d7d91\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fxrgwg" Jan 26 17:10:53 crc kubenswrapper[4856]: I0126 17:10:53.857324 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tb25l\" (UniqueName: \"kubernetes.io/projected/7105e655-ab8e-4fc0-b205-0bafaa6d7d91-kube-api-access-tb25l\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fxrgwg\" (UID: \"7105e655-ab8e-4fc0-b205-0bafaa6d7d91\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fxrgwg" Jan 26 17:10:53 crc kubenswrapper[4856]: I0126 17:10:53.857383 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7105e655-ab8e-4fc0-b205-0bafaa6d7d91-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fxrgwg\" (UID: \"7105e655-ab8e-4fc0-b205-0bafaa6d7d91\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fxrgwg" Jan 26 17:10:53 crc kubenswrapper[4856]: I0126 17:10:53.857436 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7105e655-ab8e-4fc0-b205-0bafaa6d7d91-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fxrgwg\" (UID: \"7105e655-ab8e-4fc0-b205-0bafaa6d7d91\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fxrgwg" Jan 26 17:10:53 crc kubenswrapper[4856]: I0126 17:10:53.857942 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7105e655-ab8e-4fc0-b205-0bafaa6d7d91-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fxrgwg\" (UID: \"7105e655-ab8e-4fc0-b205-0bafaa6d7d91\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fxrgwg" Jan 26 17:10:53 crc kubenswrapper[4856]: I0126 17:10:53.858614 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7105e655-ab8e-4fc0-b205-0bafaa6d7d91-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fxrgwg\" (UID: \"7105e655-ab8e-4fc0-b205-0bafaa6d7d91\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fxrgwg" Jan 26 17:10:53 crc kubenswrapper[4856]: I0126 17:10:53.879681 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tb25l\" (UniqueName: \"kubernetes.io/projected/7105e655-ab8e-4fc0-b205-0bafaa6d7d91-kube-api-access-tb25l\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fxrgwg\" (UID: \"7105e655-ab8e-4fc0-b205-0bafaa6d7d91\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fxrgwg" Jan 26 17:10:53 crc kubenswrapper[4856]: I0126 17:10:53.993648 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fxrgwg" Jan 26 17:10:54 crc kubenswrapper[4856]: I0126 17:10:54.184107 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fxrgwg"] Jan 26 17:10:54 crc kubenswrapper[4856]: W0126 17:10:54.192855 4856 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7105e655_ab8e_4fc0_b205_0bafaa6d7d91.slice/crio-56b1c09b7117c32b89d576119b3e75e1f8c7130cbb907853b17119459ef4324f WatchSource:0}: Error finding container 56b1c09b7117c32b89d576119b3e75e1f8c7130cbb907853b17119459ef4324f: Status 404 returned error can't find the container with id 56b1c09b7117c32b89d576119b3e75e1f8c7130cbb907853b17119459ef4324f Jan 26 17:10:54 crc kubenswrapper[4856]: I0126 17:10:54.726844 4856 generic.go:334] "Generic (PLEG): container finished" podID="7105e655-ab8e-4fc0-b205-0bafaa6d7d91" containerID="b3bfd31dd62bcd8a06a4db75b99fa7123e1c84f28625c8d87bd297a28e6a4deb" exitCode=0 Jan 26 17:10:54 crc kubenswrapper[4856]: I0126 17:10:54.726886 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fxrgwg" event={"ID":"7105e655-ab8e-4fc0-b205-0bafaa6d7d91","Type":"ContainerDied","Data":"b3bfd31dd62bcd8a06a4db75b99fa7123e1c84f28625c8d87bd297a28e6a4deb"} Jan 26 17:10:54 crc kubenswrapper[4856]: I0126 17:10:54.726912 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fxrgwg" event={"ID":"7105e655-ab8e-4fc0-b205-0bafaa6d7d91","Type":"ContainerStarted","Data":"56b1c09b7117c32b89d576119b3e75e1f8c7130cbb907853b17119459ef4324f"} Jan 26 17:10:55 crc kubenswrapper[4856]: I0126 17:10:55.734382 4856 generic.go:334] "Generic (PLEG): container finished" podID="26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a" containerID="3ab1e58078f8b846013be46f1e8be7d2a7305d073bb29980d5f81d71ab43e80d" exitCode=0 Jan 26 17:10:55 crc kubenswrapper[4856]: I0126 17:10:55.734692 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5e25mvq" event={"ID":"26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a","Type":"ContainerDied","Data":"3ab1e58078f8b846013be46f1e8be7d2a7305d073bb29980d5f81d71ab43e80d"} Jan 26 17:10:56 crc kubenswrapper[4856]: I0126 17:10:56.748121 4856 generic.go:334] "Generic (PLEG): container finished" podID="26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a" containerID="ab5c287e0e97e2b7c9c9e2b637238f42096fcbb354e37e64c27c55ba2ae02e28" exitCode=0 Jan 26 17:10:56 crc kubenswrapper[4856]: I0126 17:10:56.748255 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5e25mvq" event={"ID":"26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a","Type":"ContainerDied","Data":"ab5c287e0e97e2b7c9c9e2b637238f42096fcbb354e37e64c27c55ba2ae02e28"} Jan 26 17:10:57 crc kubenswrapper[4856]: I0126 17:10:57.757352 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fxrgwg" event={"ID":"7105e655-ab8e-4fc0-b205-0bafaa6d7d91","Type":"ContainerStarted","Data":"058a4d4cd2beed1c9c1c61654d4e95d5a1b7ade304852c5ed1486704fbd2de57"} Jan 26 17:10:58 crc kubenswrapper[4856]: I0126 17:10:58.469022 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5e25mvq" Jan 26 17:10:58 crc kubenswrapper[4856]: I0126 17:10:58.500136 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2wfr\" (UniqueName: \"kubernetes.io/projected/26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a-kube-api-access-j2wfr\") pod \"26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a\" (UID: \"26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a\") " Jan 26 17:10:58 crc kubenswrapper[4856]: I0126 17:10:58.500216 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a-util\") pod \"26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a\" (UID: \"26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a\") " Jan 26 17:10:58 crc kubenswrapper[4856]: I0126 17:10:58.500257 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a-bundle\") pod \"26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a\" (UID: \"26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a\") " Jan 26 17:10:58 crc kubenswrapper[4856]: I0126 17:10:58.501639 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a-bundle" (OuterVolumeSpecName: "bundle") pod "26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a" (UID: "26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:10:58 crc kubenswrapper[4856]: I0126 17:10:58.544636 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a-util" (OuterVolumeSpecName: "util") pod "26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a" (UID: "26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:10:58 crc kubenswrapper[4856]: I0126 17:10:58.551405 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a-kube-api-access-j2wfr" (OuterVolumeSpecName: "kube-api-access-j2wfr") pod "26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a" (UID: "26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a"). InnerVolumeSpecName "kube-api-access-j2wfr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:10:58 crc kubenswrapper[4856]: I0126 17:10:58.601907 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2wfr\" (UniqueName: \"kubernetes.io/projected/26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a-kube-api-access-j2wfr\") on node \"crc\" DevicePath \"\"" Jan 26 17:10:58 crc kubenswrapper[4856]: I0126 17:10:58.601951 4856 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a-util\") on node \"crc\" DevicePath \"\"" Jan 26 17:10:58 crc kubenswrapper[4856]: I0126 17:10:58.601965 4856 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:10:58 crc kubenswrapper[4856]: I0126 17:10:58.766940 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5e25mvq" event={"ID":"26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a","Type":"ContainerDied","Data":"13128ab773dafbdc20c6ca5345fca45949f139a13824371598c28655f37fe918"} Jan 26 17:10:58 crc kubenswrapper[4856]: I0126 17:10:58.766974 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5e25mvq" Jan 26 17:10:58 crc kubenswrapper[4856]: I0126 17:10:58.766987 4856 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="13128ab773dafbdc20c6ca5345fca45949f139a13824371598c28655f37fe918" Jan 26 17:10:58 crc kubenswrapper[4856]: I0126 17:10:58.769253 4856 generic.go:334] "Generic (PLEG): container finished" podID="7105e655-ab8e-4fc0-b205-0bafaa6d7d91" containerID="058a4d4cd2beed1c9c1c61654d4e95d5a1b7ade304852c5ed1486704fbd2de57" exitCode=0 Jan 26 17:10:58 crc kubenswrapper[4856]: I0126 17:10:58.769303 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fxrgwg" event={"ID":"7105e655-ab8e-4fc0-b205-0bafaa6d7d91","Type":"ContainerDied","Data":"058a4d4cd2beed1c9c1c61654d4e95d5a1b7ade304852c5ed1486704fbd2de57"} Jan 26 17:10:59 crc kubenswrapper[4856]: I0126 17:10:59.540013 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ank5l6"] Jan 26 17:10:59 crc kubenswrapper[4856]: E0126 17:10:59.540331 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a" containerName="extract" Jan 26 17:10:59 crc kubenswrapper[4856]: I0126 17:10:59.540353 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a" containerName="extract" Jan 26 17:10:59 crc kubenswrapper[4856]: E0126 17:10:59.540371 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a" containerName="pull" Jan 26 17:10:59 crc kubenswrapper[4856]: I0126 17:10:59.540380 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a" containerName="pull" Jan 26 17:10:59 crc kubenswrapper[4856]: E0126 17:10:59.540402 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a" containerName="util" Jan 26 17:10:59 crc kubenswrapper[4856]: I0126 17:10:59.540413 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a" containerName="util" Jan 26 17:10:59 crc kubenswrapper[4856]: I0126 17:10:59.540554 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a" containerName="extract" Jan 26 17:10:59 crc kubenswrapper[4856]: I0126 17:10:59.541552 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ank5l6" Jan 26 17:10:59 crc kubenswrapper[4856]: I0126 17:10:59.557964 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ank5l6"] Jan 26 17:10:59 crc kubenswrapper[4856]: I0126 17:10:59.618600 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6521dc23-8f4e-452f-ae3e-167424fa3ed2-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ank5l6\" (UID: \"6521dc23-8f4e-452f-ae3e-167424fa3ed2\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ank5l6" Jan 26 17:10:59 crc kubenswrapper[4856]: I0126 17:10:59.618704 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6521dc23-8f4e-452f-ae3e-167424fa3ed2-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ank5l6\" (UID: \"6521dc23-8f4e-452f-ae3e-167424fa3ed2\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ank5l6" Jan 26 17:10:59 crc kubenswrapper[4856]: I0126 17:10:59.618738 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vksql\" (UniqueName: \"kubernetes.io/projected/6521dc23-8f4e-452f-ae3e-167424fa3ed2-kube-api-access-vksql\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ank5l6\" (UID: \"6521dc23-8f4e-452f-ae3e-167424fa3ed2\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ank5l6" Jan 26 17:10:59 crc kubenswrapper[4856]: I0126 17:10:59.720231 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6521dc23-8f4e-452f-ae3e-167424fa3ed2-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ank5l6\" (UID: \"6521dc23-8f4e-452f-ae3e-167424fa3ed2\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ank5l6" Jan 26 17:10:59 crc kubenswrapper[4856]: I0126 17:10:59.720631 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6521dc23-8f4e-452f-ae3e-167424fa3ed2-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ank5l6\" (UID: \"6521dc23-8f4e-452f-ae3e-167424fa3ed2\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ank5l6" Jan 26 17:10:59 crc kubenswrapper[4856]: I0126 17:10:59.720662 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vksql\" (UniqueName: \"kubernetes.io/projected/6521dc23-8f4e-452f-ae3e-167424fa3ed2-kube-api-access-vksql\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ank5l6\" (UID: \"6521dc23-8f4e-452f-ae3e-167424fa3ed2\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ank5l6" Jan 26 17:10:59 crc kubenswrapper[4856]: I0126 17:10:59.720983 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6521dc23-8f4e-452f-ae3e-167424fa3ed2-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ank5l6\" (UID: \"6521dc23-8f4e-452f-ae3e-167424fa3ed2\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ank5l6" Jan 26 17:10:59 crc kubenswrapper[4856]: I0126 17:10:59.721187 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6521dc23-8f4e-452f-ae3e-167424fa3ed2-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ank5l6\" (UID: \"6521dc23-8f4e-452f-ae3e-167424fa3ed2\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ank5l6" Jan 26 17:10:59 crc kubenswrapper[4856]: I0126 17:10:59.752881 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vksql\" (UniqueName: \"kubernetes.io/projected/6521dc23-8f4e-452f-ae3e-167424fa3ed2-kube-api-access-vksql\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ank5l6\" (UID: \"6521dc23-8f4e-452f-ae3e-167424fa3ed2\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ank5l6" Jan 26 17:10:59 crc kubenswrapper[4856]: I0126 17:10:59.777156 4856 generic.go:334] "Generic (PLEG): container finished" podID="7105e655-ab8e-4fc0-b205-0bafaa6d7d91" containerID="7145619c0c090acd718be2468e255969f24b82973542e62ca9c51a6b03860c3e" exitCode=0 Jan 26 17:10:59 crc kubenswrapper[4856]: I0126 17:10:59.777205 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fxrgwg" event={"ID":"7105e655-ab8e-4fc0-b205-0bafaa6d7d91","Type":"ContainerDied","Data":"7145619c0c090acd718be2468e255969f24b82973542e62ca9c51a6b03860c3e"} Jan 26 17:10:59 crc kubenswrapper[4856]: I0126 17:10:59.924504 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ank5l6" Jan 26 17:11:01 crc kubenswrapper[4856]: I0126 17:11:01.362156 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ank5l6"] Jan 26 17:11:01 crc kubenswrapper[4856]: I0126 17:11:01.534837 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fxrgwg" Jan 26 17:11:01 crc kubenswrapper[4856]: I0126 17:11:01.611640 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7105e655-ab8e-4fc0-b205-0bafaa6d7d91-util\") pod \"7105e655-ab8e-4fc0-b205-0bafaa6d7d91\" (UID: \"7105e655-ab8e-4fc0-b205-0bafaa6d7d91\") " Jan 26 17:11:01 crc kubenswrapper[4856]: I0126 17:11:01.611832 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7105e655-ab8e-4fc0-b205-0bafaa6d7d91-bundle\") pod \"7105e655-ab8e-4fc0-b205-0bafaa6d7d91\" (UID: \"7105e655-ab8e-4fc0-b205-0bafaa6d7d91\") " Jan 26 17:11:01 crc kubenswrapper[4856]: I0126 17:11:01.611876 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tb25l\" (UniqueName: \"kubernetes.io/projected/7105e655-ab8e-4fc0-b205-0bafaa6d7d91-kube-api-access-tb25l\") pod \"7105e655-ab8e-4fc0-b205-0bafaa6d7d91\" (UID: \"7105e655-ab8e-4fc0-b205-0bafaa6d7d91\") " Jan 26 17:11:01 crc kubenswrapper[4856]: I0126 17:11:01.612963 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7105e655-ab8e-4fc0-b205-0bafaa6d7d91-bundle" (OuterVolumeSpecName: "bundle") pod "7105e655-ab8e-4fc0-b205-0bafaa6d7d91" (UID: "7105e655-ab8e-4fc0-b205-0bafaa6d7d91"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:11:01 crc kubenswrapper[4856]: I0126 17:11:01.632784 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7105e655-ab8e-4fc0-b205-0bafaa6d7d91-kube-api-access-tb25l" (OuterVolumeSpecName: "kube-api-access-tb25l") pod "7105e655-ab8e-4fc0-b205-0bafaa6d7d91" (UID: "7105e655-ab8e-4fc0-b205-0bafaa6d7d91"). InnerVolumeSpecName "kube-api-access-tb25l". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:11:01 crc kubenswrapper[4856]: I0126 17:11:01.636673 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7105e655-ab8e-4fc0-b205-0bafaa6d7d91-util" (OuterVolumeSpecName: "util") pod "7105e655-ab8e-4fc0-b205-0bafaa6d7d91" (UID: "7105e655-ab8e-4fc0-b205-0bafaa6d7d91"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:11:01 crc kubenswrapper[4856]: I0126 17:11:01.713766 4856 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7105e655-ab8e-4fc0-b205-0bafaa6d7d91-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:01 crc kubenswrapper[4856]: I0126 17:11:01.714118 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tb25l\" (UniqueName: \"kubernetes.io/projected/7105e655-ab8e-4fc0-b205-0bafaa6d7d91-kube-api-access-tb25l\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:01 crc kubenswrapper[4856]: I0126 17:11:01.714131 4856 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7105e655-ab8e-4fc0-b205-0bafaa6d7d91-util\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:01 crc kubenswrapper[4856]: I0126 17:11:01.874509 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ank5l6" event={"ID":"6521dc23-8f4e-452f-ae3e-167424fa3ed2","Type":"ContainerStarted","Data":"2a431074e034cf10e4752b83a72e67b08ab250c4805b0ef25b26d6818d7e9e5d"} Jan 26 17:11:01 crc kubenswrapper[4856]: I0126 17:11:01.876801 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fxrgwg" event={"ID":"7105e655-ab8e-4fc0-b205-0bafaa6d7d91","Type":"ContainerDied","Data":"56b1c09b7117c32b89d576119b3e75e1f8c7130cbb907853b17119459ef4324f"} Jan 26 17:11:01 crc kubenswrapper[4856]: I0126 17:11:01.876865 4856 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="56b1c09b7117c32b89d576119b3e75e1f8c7130cbb907853b17119459ef4324f" Jan 26 17:11:01 crc kubenswrapper[4856]: I0126 17:11:01.876862 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fxrgwg" Jan 26 17:11:02 crc kubenswrapper[4856]: I0126 17:11:02.963652 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ank5l6" event={"ID":"6521dc23-8f4e-452f-ae3e-167424fa3ed2","Type":"ContainerStarted","Data":"c153773f3fa63e0d5a49b944e935649f372715f37305b42c32405d4d2a56f4ad"} Jan 26 17:11:03 crc kubenswrapper[4856]: I0126 17:11:03.971455 4856 generic.go:334] "Generic (PLEG): container finished" podID="6521dc23-8f4e-452f-ae3e-167424fa3ed2" containerID="c153773f3fa63e0d5a49b944e935649f372715f37305b42c32405d4d2a56f4ad" exitCode=0 Jan 26 17:11:03 crc kubenswrapper[4856]: I0126 17:11:03.971603 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ank5l6" event={"ID":"6521dc23-8f4e-452f-ae3e-167424fa3ed2","Type":"ContainerDied","Data":"c153773f3fa63e0d5a49b944e935649f372715f37305b42c32405d4d2a56f4ad"} Jan 26 17:11:05 crc kubenswrapper[4856]: I0126 17:11:05.736241 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-cq2gx"] Jan 26 17:11:05 crc kubenswrapper[4856]: E0126 17:11:05.736830 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7105e655-ab8e-4fc0-b205-0bafaa6d7d91" containerName="pull" Jan 26 17:11:05 crc kubenswrapper[4856]: I0126 17:11:05.736845 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="7105e655-ab8e-4fc0-b205-0bafaa6d7d91" containerName="pull" Jan 26 17:11:05 crc kubenswrapper[4856]: E0126 17:11:05.736862 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7105e655-ab8e-4fc0-b205-0bafaa6d7d91" containerName="extract" Jan 26 17:11:05 crc kubenswrapper[4856]: I0126 17:11:05.736868 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="7105e655-ab8e-4fc0-b205-0bafaa6d7d91" containerName="extract" Jan 26 17:11:05 crc kubenswrapper[4856]: E0126 17:11:05.736878 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7105e655-ab8e-4fc0-b205-0bafaa6d7d91" containerName="util" Jan 26 17:11:05 crc kubenswrapper[4856]: I0126 17:11:05.736883 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="7105e655-ab8e-4fc0-b205-0bafaa6d7d91" containerName="util" Jan 26 17:11:05 crc kubenswrapper[4856]: I0126 17:11:05.736991 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="7105e655-ab8e-4fc0-b205-0bafaa6d7d91" containerName="extract" Jan 26 17:11:05 crc kubenswrapper[4856]: I0126 17:11:05.737353 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-cq2gx" Jan 26 17:11:05 crc kubenswrapper[4856]: I0126 17:11:05.744644 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Jan 26 17:11:05 crc kubenswrapper[4856]: I0126 17:11:05.751219 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-k99fx" Jan 26 17:11:05 crc kubenswrapper[4856]: I0126 17:11:05.751272 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Jan 26 17:11:05 crc kubenswrapper[4856]: I0126 17:11:05.762068 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-cq2gx"] Jan 26 17:11:05 crc kubenswrapper[4856]: I0126 17:11:05.848407 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmwsq\" (UniqueName: \"kubernetes.io/projected/e31d2d53-8992-45e3-98aa-24ea73236248-kube-api-access-bmwsq\") pod \"obo-prometheus-operator-68bc856cb9-cq2gx\" (UID: \"e31d2d53-8992-45e3-98aa-24ea73236248\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-cq2gx" Jan 26 17:11:05 crc kubenswrapper[4856]: I0126 17:11:05.889863 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-68d9fdc4dd-jbq25"] Jan 26 17:11:05 crc kubenswrapper[4856]: I0126 17:11:05.890971 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d9fdc4dd-jbq25" Jan 26 17:11:05 crc kubenswrapper[4856]: I0126 17:11:05.893443 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Jan 26 17:11:05 crc kubenswrapper[4856]: I0126 17:11:05.896975 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-lmkrv" Jan 26 17:11:05 crc kubenswrapper[4856]: I0126 17:11:05.903205 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-68d9fdc4dd-cf7wn"] Jan 26 17:11:05 crc kubenswrapper[4856]: I0126 17:11:05.904100 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d9fdc4dd-cf7wn" Jan 26 17:11:05 crc kubenswrapper[4856]: I0126 17:11:05.915624 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-68d9fdc4dd-jbq25"] Jan 26 17:11:05 crc kubenswrapper[4856]: I0126 17:11:05.925644 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-68d9fdc4dd-cf7wn"] Jan 26 17:11:05 crc kubenswrapper[4856]: I0126 17:11:05.953626 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmwsq\" (UniqueName: \"kubernetes.io/projected/e31d2d53-8992-45e3-98aa-24ea73236248-kube-api-access-bmwsq\") pod \"obo-prometheus-operator-68bc856cb9-cq2gx\" (UID: \"e31d2d53-8992-45e3-98aa-24ea73236248\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-cq2gx" Jan 26 17:11:05 crc kubenswrapper[4856]: I0126 17:11:05.953700 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7c88687f-1304-4709-b148-a196f0d0190d-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-68d9fdc4dd-cf7wn\" (UID: \"7c88687f-1304-4709-b148-a196f0d0190d\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d9fdc4dd-cf7wn" Jan 26 17:11:05 crc kubenswrapper[4856]: I0126 17:11:05.953802 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/766f50ba-0751-4f25-a6db-3b7195e72f55-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-68d9fdc4dd-jbq25\" (UID: \"766f50ba-0751-4f25-a6db-3b7195e72f55\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d9fdc4dd-jbq25" Jan 26 17:11:05 crc kubenswrapper[4856]: I0126 17:11:05.953864 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/766f50ba-0751-4f25-a6db-3b7195e72f55-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-68d9fdc4dd-jbq25\" (UID: \"766f50ba-0751-4f25-a6db-3b7195e72f55\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d9fdc4dd-jbq25" Jan 26 17:11:05 crc kubenswrapper[4856]: I0126 17:11:05.953896 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7c88687f-1304-4709-b148-a196f0d0190d-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-68d9fdc4dd-cf7wn\" (UID: \"7c88687f-1304-4709-b148-a196f0d0190d\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d9fdc4dd-cf7wn" Jan 26 17:11:05 crc kubenswrapper[4856]: I0126 17:11:05.978852 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmwsq\" (UniqueName: \"kubernetes.io/projected/e31d2d53-8992-45e3-98aa-24ea73236248-kube-api-access-bmwsq\") pod \"obo-prometheus-operator-68bc856cb9-cq2gx\" (UID: \"e31d2d53-8992-45e3-98aa-24ea73236248\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-cq2gx" Jan 26 17:11:06 crc kubenswrapper[4856]: I0126 17:11:06.055473 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/766f50ba-0751-4f25-a6db-3b7195e72f55-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-68d9fdc4dd-jbq25\" (UID: \"766f50ba-0751-4f25-a6db-3b7195e72f55\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d9fdc4dd-jbq25" Jan 26 17:11:06 crc kubenswrapper[4856]: I0126 17:11:06.055534 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7c88687f-1304-4709-b148-a196f0d0190d-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-68d9fdc4dd-cf7wn\" (UID: \"7c88687f-1304-4709-b148-a196f0d0190d\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d9fdc4dd-cf7wn" Jan 26 17:11:06 crc kubenswrapper[4856]: I0126 17:11:06.055569 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7c88687f-1304-4709-b148-a196f0d0190d-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-68d9fdc4dd-cf7wn\" (UID: \"7c88687f-1304-4709-b148-a196f0d0190d\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d9fdc4dd-cf7wn" Jan 26 17:11:06 crc kubenswrapper[4856]: I0126 17:11:06.055622 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/766f50ba-0751-4f25-a6db-3b7195e72f55-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-68d9fdc4dd-jbq25\" (UID: \"766f50ba-0751-4f25-a6db-3b7195e72f55\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d9fdc4dd-jbq25" Jan 26 17:11:06 crc kubenswrapper[4856]: I0126 17:11:06.056149 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-cq2gx" Jan 26 17:11:06 crc kubenswrapper[4856]: I0126 17:11:06.059761 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7c88687f-1304-4709-b148-a196f0d0190d-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-68d9fdc4dd-cf7wn\" (UID: \"7c88687f-1304-4709-b148-a196f0d0190d\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d9fdc4dd-cf7wn" Jan 26 17:11:06 crc kubenswrapper[4856]: I0126 17:11:06.060207 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7c88687f-1304-4709-b148-a196f0d0190d-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-68d9fdc4dd-cf7wn\" (UID: \"7c88687f-1304-4709-b148-a196f0d0190d\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d9fdc4dd-cf7wn" Jan 26 17:11:06 crc kubenswrapper[4856]: I0126 17:11:06.079102 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/766f50ba-0751-4f25-a6db-3b7195e72f55-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-68d9fdc4dd-jbq25\" (UID: \"766f50ba-0751-4f25-a6db-3b7195e72f55\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d9fdc4dd-jbq25" Jan 26 17:11:06 crc kubenswrapper[4856]: I0126 17:11:06.091182 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/766f50ba-0751-4f25-a6db-3b7195e72f55-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-68d9fdc4dd-jbq25\" (UID: \"766f50ba-0751-4f25-a6db-3b7195e72f55\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d9fdc4dd-jbq25" Jan 26 17:11:06 crc kubenswrapper[4856]: I0126 17:11:06.165899 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-fpn2h"] Jan 26 17:11:06 crc kubenswrapper[4856]: I0126 17:11:06.166667 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-fpn2h" Jan 26 17:11:06 crc kubenswrapper[4856]: I0126 17:11:06.171906 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Jan 26 17:11:06 crc kubenswrapper[4856]: I0126 17:11:06.174220 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-8x6vt" Jan 26 17:11:06 crc kubenswrapper[4856]: I0126 17:11:06.182405 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-fpn2h"] Jan 26 17:11:06 crc kubenswrapper[4856]: I0126 17:11:06.205427 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d9fdc4dd-jbq25" Jan 26 17:11:06 crc kubenswrapper[4856]: I0126 17:11:06.217782 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d9fdc4dd-cf7wn" Jan 26 17:11:06 crc kubenswrapper[4856]: I0126 17:11:06.258799 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/a4ae7646-2afb-4ada-b8a4-d20a69f87949-observability-operator-tls\") pod \"observability-operator-59bdc8b94-fpn2h\" (UID: \"a4ae7646-2afb-4ada-b8a4-d20a69f87949\") " pod="openshift-operators/observability-operator-59bdc8b94-fpn2h" Jan 26 17:11:06 crc kubenswrapper[4856]: I0126 17:11:06.258867 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2ptr\" (UniqueName: \"kubernetes.io/projected/a4ae7646-2afb-4ada-b8a4-d20a69f87949-kube-api-access-n2ptr\") pod \"observability-operator-59bdc8b94-fpn2h\" (UID: \"a4ae7646-2afb-4ada-b8a4-d20a69f87949\") " pod="openshift-operators/observability-operator-59bdc8b94-fpn2h" Jan 26 17:11:06 crc kubenswrapper[4856]: I0126 17:11:06.304954 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-5bmfp"] Jan 26 17:11:06 crc kubenswrapper[4856]: I0126 17:11:06.306618 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-5bmfp" Jan 26 17:11:06 crc kubenswrapper[4856]: I0126 17:11:06.311124 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-6jlcg" Jan 26 17:11:06 crc kubenswrapper[4856]: I0126 17:11:06.349743 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-5bmfp"] Jan 26 17:11:06 crc kubenswrapper[4856]: I0126 17:11:06.427478 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmjhr\" (UniqueName: \"kubernetes.io/projected/bd7597f2-d44b-4e1b-ac60-b409985e3351-kube-api-access-tmjhr\") pod \"perses-operator-5bf474d74f-5bmfp\" (UID: \"bd7597f2-d44b-4e1b-ac60-b409985e3351\") " pod="openshift-operators/perses-operator-5bf474d74f-5bmfp" Jan 26 17:11:06 crc kubenswrapper[4856]: I0126 17:11:06.427616 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/bd7597f2-d44b-4e1b-ac60-b409985e3351-openshift-service-ca\") pod \"perses-operator-5bf474d74f-5bmfp\" (UID: \"bd7597f2-d44b-4e1b-ac60-b409985e3351\") " pod="openshift-operators/perses-operator-5bf474d74f-5bmfp" Jan 26 17:11:06 crc kubenswrapper[4856]: I0126 17:11:06.427717 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/a4ae7646-2afb-4ada-b8a4-d20a69f87949-observability-operator-tls\") pod \"observability-operator-59bdc8b94-fpn2h\" (UID: \"a4ae7646-2afb-4ada-b8a4-d20a69f87949\") " pod="openshift-operators/observability-operator-59bdc8b94-fpn2h" Jan 26 17:11:06 crc kubenswrapper[4856]: I0126 17:11:06.427808 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2ptr\" (UniqueName: \"kubernetes.io/projected/a4ae7646-2afb-4ada-b8a4-d20a69f87949-kube-api-access-n2ptr\") pod \"observability-operator-59bdc8b94-fpn2h\" (UID: \"a4ae7646-2afb-4ada-b8a4-d20a69f87949\") " pod="openshift-operators/observability-operator-59bdc8b94-fpn2h" Jan 26 17:11:06 crc kubenswrapper[4856]: I0126 17:11:06.447779 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/a4ae7646-2afb-4ada-b8a4-d20a69f87949-observability-operator-tls\") pod \"observability-operator-59bdc8b94-fpn2h\" (UID: \"a4ae7646-2afb-4ada-b8a4-d20a69f87949\") " pod="openshift-operators/observability-operator-59bdc8b94-fpn2h" Jan 26 17:11:06 crc kubenswrapper[4856]: I0126 17:11:06.450429 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2ptr\" (UniqueName: \"kubernetes.io/projected/a4ae7646-2afb-4ada-b8a4-d20a69f87949-kube-api-access-n2ptr\") pod \"observability-operator-59bdc8b94-fpn2h\" (UID: \"a4ae7646-2afb-4ada-b8a4-d20a69f87949\") " pod="openshift-operators/observability-operator-59bdc8b94-fpn2h" Jan 26 17:11:06 crc kubenswrapper[4856]: I0126 17:11:06.500476 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-fpn2h" Jan 26 17:11:06 crc kubenswrapper[4856]: I0126 17:11:06.533234 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmjhr\" (UniqueName: \"kubernetes.io/projected/bd7597f2-d44b-4e1b-ac60-b409985e3351-kube-api-access-tmjhr\") pod \"perses-operator-5bf474d74f-5bmfp\" (UID: \"bd7597f2-d44b-4e1b-ac60-b409985e3351\") " pod="openshift-operators/perses-operator-5bf474d74f-5bmfp" Jan 26 17:11:06 crc kubenswrapper[4856]: I0126 17:11:06.533313 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/bd7597f2-d44b-4e1b-ac60-b409985e3351-openshift-service-ca\") pod \"perses-operator-5bf474d74f-5bmfp\" (UID: \"bd7597f2-d44b-4e1b-ac60-b409985e3351\") " pod="openshift-operators/perses-operator-5bf474d74f-5bmfp" Jan 26 17:11:06 crc kubenswrapper[4856]: I0126 17:11:06.535188 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/bd7597f2-d44b-4e1b-ac60-b409985e3351-openshift-service-ca\") pod \"perses-operator-5bf474d74f-5bmfp\" (UID: \"bd7597f2-d44b-4e1b-ac60-b409985e3351\") " pod="openshift-operators/perses-operator-5bf474d74f-5bmfp" Jan 26 17:11:06 crc kubenswrapper[4856]: I0126 17:11:06.610388 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmjhr\" (UniqueName: \"kubernetes.io/projected/bd7597f2-d44b-4e1b-ac60-b409985e3351-kube-api-access-tmjhr\") pod \"perses-operator-5bf474d74f-5bmfp\" (UID: \"bd7597f2-d44b-4e1b-ac60-b409985e3351\") " pod="openshift-operators/perses-operator-5bf474d74f-5bmfp" Jan 26 17:11:06 crc kubenswrapper[4856]: I0126 17:11:06.686708 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-cq2gx"] Jan 26 17:11:06 crc kubenswrapper[4856]: I0126 17:11:06.691050 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-5bmfp" Jan 26 17:11:06 crc kubenswrapper[4856]: I0126 17:11:06.837065 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-68d9fdc4dd-cf7wn"] Jan 26 17:11:06 crc kubenswrapper[4856]: W0126 17:11:06.860801 4856 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c88687f_1304_4709_b148_a196f0d0190d.slice/crio-3add66ed30f57595204f43560d849f6ef5b5e72b9812deeb4561fc229fac034a WatchSource:0}: Error finding container 3add66ed30f57595204f43560d849f6ef5b5e72b9812deeb4561fc229fac034a: Status 404 returned error can't find the container with id 3add66ed30f57595204f43560d849f6ef5b5e72b9812deeb4561fc229fac034a Jan 26 17:11:06 crc kubenswrapper[4856]: I0126 17:11:06.932901 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-fpn2h"] Jan 26 17:11:06 crc kubenswrapper[4856]: W0126 17:11:06.938469 4856 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda4ae7646_2afb_4ada_b8a4_d20a69f87949.slice/crio-0d3567099a83e901da442c5a280e90fc2283599dea6ed1bb6faac6fbd1659711 WatchSource:0}: Error finding container 0d3567099a83e901da442c5a280e90fc2283599dea6ed1bb6faac6fbd1659711: Status 404 returned error can't find the container with id 0d3567099a83e901da442c5a280e90fc2283599dea6ed1bb6faac6fbd1659711 Jan 26 17:11:06 crc kubenswrapper[4856]: I0126 17:11:06.958300 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-68d9fdc4dd-jbq25"] Jan 26 17:11:06 crc kubenswrapper[4856]: I0126 17:11:06.972231 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-5bmfp"] Jan 26 17:11:06 crc kubenswrapper[4856]: W0126 17:11:06.974670 4856 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod766f50ba_0751_4f25_a6db_3b7195e72f55.slice/crio-058bea58322e620b766ee5384d671712d335f6e308c7fab9e9134a7e4f0b21f4 WatchSource:0}: Error finding container 058bea58322e620b766ee5384d671712d335f6e308c7fab9e9134a7e4f0b21f4: Status 404 returned error can't find the container with id 058bea58322e620b766ee5384d671712d335f6e308c7fab9e9134a7e4f0b21f4 Jan 26 17:11:06 crc kubenswrapper[4856]: W0126 17:11:06.978655 4856 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbd7597f2_d44b_4e1b_ac60_b409985e3351.slice/crio-de73b797d25e7aab5cc4db5082d70b93dde8cc46381630d4b1bb2cc17b926363 WatchSource:0}: Error finding container de73b797d25e7aab5cc4db5082d70b93dde8cc46381630d4b1bb2cc17b926363: Status 404 returned error can't find the container with id de73b797d25e7aab5cc4db5082d70b93dde8cc46381630d4b1bb2cc17b926363 Jan 26 17:11:07 crc kubenswrapper[4856]: I0126 17:11:07.042126 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-cq2gx" event={"ID":"e31d2d53-8992-45e3-98aa-24ea73236248","Type":"ContainerStarted","Data":"89868836e252b26754a35e23094de18e5e17220691fdb6335046087453e4be01"} Jan 26 17:11:07 crc kubenswrapper[4856]: I0126 17:11:07.048616 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d9fdc4dd-cf7wn" event={"ID":"7c88687f-1304-4709-b148-a196f0d0190d","Type":"ContainerStarted","Data":"3add66ed30f57595204f43560d849f6ef5b5e72b9812deeb4561fc229fac034a"} Jan 26 17:11:07 crc kubenswrapper[4856]: I0126 17:11:07.049884 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-fpn2h" event={"ID":"a4ae7646-2afb-4ada-b8a4-d20a69f87949","Type":"ContainerStarted","Data":"0d3567099a83e901da442c5a280e90fc2283599dea6ed1bb6faac6fbd1659711"} Jan 26 17:11:07 crc kubenswrapper[4856]: I0126 17:11:07.051171 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-5bmfp" event={"ID":"bd7597f2-d44b-4e1b-ac60-b409985e3351","Type":"ContainerStarted","Data":"de73b797d25e7aab5cc4db5082d70b93dde8cc46381630d4b1bb2cc17b926363"} Jan 26 17:11:07 crc kubenswrapper[4856]: I0126 17:11:07.052477 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d9fdc4dd-jbq25" event={"ID":"766f50ba-0751-4f25-a6db-3b7195e72f55","Type":"ContainerStarted","Data":"058bea58322e620b766ee5384d671712d335f6e308c7fab9e9134a7e4f0b21f4"} Jan 26 17:11:08 crc kubenswrapper[4856]: I0126 17:11:08.599905 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/elastic-operator-bf765cf6c-gbst6"] Jan 26 17:11:08 crc kubenswrapper[4856]: I0126 17:11:08.600948 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elastic-operator-bf765cf6c-gbst6" Jan 26 17:11:08 crc kubenswrapper[4856]: I0126 17:11:08.602881 4856 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elastic-operator-service-cert" Jan 26 17:11:08 crc kubenswrapper[4856]: I0126 17:11:08.603506 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"openshift-service-ca.crt" Jan 26 17:11:08 crc kubenswrapper[4856]: I0126 17:11:08.603783 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"kube-root-ca.crt" Jan 26 17:11:08 crc kubenswrapper[4856]: I0126 17:11:08.603802 4856 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elastic-operator-dockercfg-bxj5j" Jan 26 17:11:08 crc kubenswrapper[4856]: I0126 17:11:08.608641 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elastic-operator-bf765cf6c-gbst6"] Jan 26 17:11:08 crc kubenswrapper[4856]: I0126 17:11:08.666194 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9988655e-7b1f-443e-a102-68665719162a-apiservice-cert\") pod \"elastic-operator-bf765cf6c-gbst6\" (UID: \"9988655e-7b1f-443e-a102-68665719162a\") " pod="service-telemetry/elastic-operator-bf765cf6c-gbst6" Jan 26 17:11:08 crc kubenswrapper[4856]: I0126 17:11:08.666295 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7wld\" (UniqueName: \"kubernetes.io/projected/9988655e-7b1f-443e-a102-68665719162a-kube-api-access-z7wld\") pod \"elastic-operator-bf765cf6c-gbst6\" (UID: \"9988655e-7b1f-443e-a102-68665719162a\") " pod="service-telemetry/elastic-operator-bf765cf6c-gbst6" Jan 26 17:11:08 crc kubenswrapper[4856]: I0126 17:11:08.666328 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9988655e-7b1f-443e-a102-68665719162a-webhook-cert\") pod \"elastic-operator-bf765cf6c-gbst6\" (UID: \"9988655e-7b1f-443e-a102-68665719162a\") " pod="service-telemetry/elastic-operator-bf765cf6c-gbst6" Jan 26 17:11:08 crc kubenswrapper[4856]: I0126 17:11:08.767607 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7wld\" (UniqueName: \"kubernetes.io/projected/9988655e-7b1f-443e-a102-68665719162a-kube-api-access-z7wld\") pod \"elastic-operator-bf765cf6c-gbst6\" (UID: \"9988655e-7b1f-443e-a102-68665719162a\") " pod="service-telemetry/elastic-operator-bf765cf6c-gbst6" Jan 26 17:11:08 crc kubenswrapper[4856]: I0126 17:11:08.767664 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9988655e-7b1f-443e-a102-68665719162a-webhook-cert\") pod \"elastic-operator-bf765cf6c-gbst6\" (UID: \"9988655e-7b1f-443e-a102-68665719162a\") " pod="service-telemetry/elastic-operator-bf765cf6c-gbst6" Jan 26 17:11:08 crc kubenswrapper[4856]: I0126 17:11:08.767731 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9988655e-7b1f-443e-a102-68665719162a-apiservice-cert\") pod \"elastic-operator-bf765cf6c-gbst6\" (UID: \"9988655e-7b1f-443e-a102-68665719162a\") " pod="service-telemetry/elastic-operator-bf765cf6c-gbst6" Jan 26 17:11:08 crc kubenswrapper[4856]: I0126 17:11:08.776830 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9988655e-7b1f-443e-a102-68665719162a-apiservice-cert\") pod \"elastic-operator-bf765cf6c-gbst6\" (UID: \"9988655e-7b1f-443e-a102-68665719162a\") " pod="service-telemetry/elastic-operator-bf765cf6c-gbst6" Jan 26 17:11:08 crc kubenswrapper[4856]: I0126 17:11:08.777411 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9988655e-7b1f-443e-a102-68665719162a-webhook-cert\") pod \"elastic-operator-bf765cf6c-gbst6\" (UID: \"9988655e-7b1f-443e-a102-68665719162a\") " pod="service-telemetry/elastic-operator-bf765cf6c-gbst6" Jan 26 17:11:08 crc kubenswrapper[4856]: I0126 17:11:08.797886 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7wld\" (UniqueName: \"kubernetes.io/projected/9988655e-7b1f-443e-a102-68665719162a-kube-api-access-z7wld\") pod \"elastic-operator-bf765cf6c-gbst6\" (UID: \"9988655e-7b1f-443e-a102-68665719162a\") " pod="service-telemetry/elastic-operator-bf765cf6c-gbst6" Jan 26 17:11:08 crc kubenswrapper[4856]: I0126 17:11:08.932196 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elastic-operator-bf765cf6c-gbst6" Jan 26 17:11:11 crc kubenswrapper[4856]: I0126 17:11:11.556924 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elastic-operator-bf765cf6c-gbst6"] Jan 26 17:11:12 crc kubenswrapper[4856]: I0126 17:11:12.097874 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elastic-operator-bf765cf6c-gbst6" event={"ID":"9988655e-7b1f-443e-a102-68665719162a","Type":"ContainerStarted","Data":"233316c791f13f4d4271301bb5da38f318ff55fa2e90ea73e86d39aaa58b095b"} Jan 26 17:11:12 crc kubenswrapper[4856]: I0126 17:11:12.102405 4856 generic.go:334] "Generic (PLEG): container finished" podID="6521dc23-8f4e-452f-ae3e-167424fa3ed2" containerID="bf1c283db10cced461f31388c1dde7855a66c215b11b413948361c0c6d4c1c16" exitCode=0 Jan 26 17:11:12 crc kubenswrapper[4856]: I0126 17:11:12.102447 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ank5l6" event={"ID":"6521dc23-8f4e-452f-ae3e-167424fa3ed2","Type":"ContainerDied","Data":"bf1c283db10cced461f31388c1dde7855a66c215b11b413948361c0c6d4c1c16"} Jan 26 17:11:12 crc kubenswrapper[4856]: I0126 17:11:12.362975 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-k8bc4"] Jan 26 17:11:12 crc kubenswrapper[4856]: I0126 17:11:12.364221 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/interconnect-operator-5bb49f789d-k8bc4" Jan 26 17:11:12 crc kubenswrapper[4856]: I0126 17:11:12.368018 4856 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"interconnect-operator-dockercfg-bffzw" Jan 26 17:11:12 crc kubenswrapper[4856]: I0126 17:11:12.381066 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-k8bc4"] Jan 26 17:11:12 crc kubenswrapper[4856]: I0126 17:11:12.437629 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhz7k\" (UniqueName: \"kubernetes.io/projected/15358471-0b96-4792-aaaa-433823f9ed88-kube-api-access-qhz7k\") pod \"interconnect-operator-5bb49f789d-k8bc4\" (UID: \"15358471-0b96-4792-aaaa-433823f9ed88\") " pod="service-telemetry/interconnect-operator-5bb49f789d-k8bc4" Jan 26 17:11:12 crc kubenswrapper[4856]: I0126 17:11:12.538871 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhz7k\" (UniqueName: \"kubernetes.io/projected/15358471-0b96-4792-aaaa-433823f9ed88-kube-api-access-qhz7k\") pod \"interconnect-operator-5bb49f789d-k8bc4\" (UID: \"15358471-0b96-4792-aaaa-433823f9ed88\") " pod="service-telemetry/interconnect-operator-5bb49f789d-k8bc4" Jan 26 17:11:12 crc kubenswrapper[4856]: I0126 17:11:12.561907 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhz7k\" (UniqueName: \"kubernetes.io/projected/15358471-0b96-4792-aaaa-433823f9ed88-kube-api-access-qhz7k\") pod \"interconnect-operator-5bb49f789d-k8bc4\" (UID: \"15358471-0b96-4792-aaaa-433823f9ed88\") " pod="service-telemetry/interconnect-operator-5bb49f789d-k8bc4" Jan 26 17:11:12 crc kubenswrapper[4856]: I0126 17:11:12.690258 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/interconnect-operator-5bb49f789d-k8bc4" Jan 26 17:11:13 crc kubenswrapper[4856]: I0126 17:11:13.083746 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-k8bc4"] Jan 26 17:11:13 crc kubenswrapper[4856]: I0126 17:11:13.111044 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/interconnect-operator-5bb49f789d-k8bc4" event={"ID":"15358471-0b96-4792-aaaa-433823f9ed88","Type":"ContainerStarted","Data":"53478756d067fa5ddec5eb67ee04c2f13aedb648f04c0c046a78f3cca4c0a970"} Jan 26 17:11:13 crc kubenswrapper[4856]: I0126 17:11:13.121809 4856 generic.go:334] "Generic (PLEG): container finished" podID="6521dc23-8f4e-452f-ae3e-167424fa3ed2" containerID="1b9c15a7af0a5cddb04adabfe164e630c584bd42ccb27b1da654ea8393613616" exitCode=0 Jan 26 17:11:13 crc kubenswrapper[4856]: I0126 17:11:13.121891 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ank5l6" event={"ID":"6521dc23-8f4e-452f-ae3e-167424fa3ed2","Type":"ContainerDied","Data":"1b9c15a7af0a5cddb04adabfe164e630c584bd42ccb27b1da654ea8393613616"} Jan 26 17:11:19 crc kubenswrapper[4856]: I0126 17:11:19.824934 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ank5l6" Jan 26 17:11:19 crc kubenswrapper[4856]: I0126 17:11:19.872556 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vksql\" (UniqueName: \"kubernetes.io/projected/6521dc23-8f4e-452f-ae3e-167424fa3ed2-kube-api-access-vksql\") pod \"6521dc23-8f4e-452f-ae3e-167424fa3ed2\" (UID: \"6521dc23-8f4e-452f-ae3e-167424fa3ed2\") " Jan 26 17:11:19 crc kubenswrapper[4856]: I0126 17:11:19.872711 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6521dc23-8f4e-452f-ae3e-167424fa3ed2-util\") pod \"6521dc23-8f4e-452f-ae3e-167424fa3ed2\" (UID: \"6521dc23-8f4e-452f-ae3e-167424fa3ed2\") " Jan 26 17:11:19 crc kubenswrapper[4856]: I0126 17:11:19.872736 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6521dc23-8f4e-452f-ae3e-167424fa3ed2-bundle\") pod \"6521dc23-8f4e-452f-ae3e-167424fa3ed2\" (UID: \"6521dc23-8f4e-452f-ae3e-167424fa3ed2\") " Jan 26 17:11:19 crc kubenswrapper[4856]: I0126 17:11:19.874364 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6521dc23-8f4e-452f-ae3e-167424fa3ed2-bundle" (OuterVolumeSpecName: "bundle") pod "6521dc23-8f4e-452f-ae3e-167424fa3ed2" (UID: "6521dc23-8f4e-452f-ae3e-167424fa3ed2"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:11:19 crc kubenswrapper[4856]: I0126 17:11:19.880311 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6521dc23-8f4e-452f-ae3e-167424fa3ed2-kube-api-access-vksql" (OuterVolumeSpecName: "kube-api-access-vksql") pod "6521dc23-8f4e-452f-ae3e-167424fa3ed2" (UID: "6521dc23-8f4e-452f-ae3e-167424fa3ed2"). InnerVolumeSpecName "kube-api-access-vksql". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:11:19 crc kubenswrapper[4856]: I0126 17:11:19.886074 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6521dc23-8f4e-452f-ae3e-167424fa3ed2-util" (OuterVolumeSpecName: "util") pod "6521dc23-8f4e-452f-ae3e-167424fa3ed2" (UID: "6521dc23-8f4e-452f-ae3e-167424fa3ed2"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:11:19 crc kubenswrapper[4856]: I0126 17:11:19.973747 4856 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6521dc23-8f4e-452f-ae3e-167424fa3ed2-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:19 crc kubenswrapper[4856]: I0126 17:11:19.973786 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vksql\" (UniqueName: \"kubernetes.io/projected/6521dc23-8f4e-452f-ae3e-167424fa3ed2-kube-api-access-vksql\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:19 crc kubenswrapper[4856]: I0126 17:11:19.973807 4856 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6521dc23-8f4e-452f-ae3e-167424fa3ed2-util\") on node \"crc\" DevicePath \"\"" Jan 26 17:11:20 crc kubenswrapper[4856]: I0126 17:11:20.209034 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ank5l6" event={"ID":"6521dc23-8f4e-452f-ae3e-167424fa3ed2","Type":"ContainerDied","Data":"2a431074e034cf10e4752b83a72e67b08ab250c4805b0ef25b26d6818d7e9e5d"} Jan 26 17:11:20 crc kubenswrapper[4856]: I0126 17:11:20.209090 4856 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a431074e034cf10e4752b83a72e67b08ab250c4805b0ef25b26d6818d7e9e5d" Jan 26 17:11:20 crc kubenswrapper[4856]: I0126 17:11:20.209183 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ank5l6" Jan 26 17:11:25 crc kubenswrapper[4856]: I0126 17:11:25.650308 4856 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 26 17:11:28 crc kubenswrapper[4856]: E0126 17:11:28.244784 4856 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/obo-prometheus-rhel9-operator@sha256:e7e5f4c5e8ab0ba298ef0295a7137d438a42eb177d9322212cde6ba8f367912a" Jan 26 17:11:28 crc kubenswrapper[4856]: E0126 17:11:28.245307 4856 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:prometheus-operator,Image:registry.redhat.io/cluster-observability-operator/obo-prometheus-rhel9-operator@sha256:e7e5f4c5e8ab0ba298ef0295a7137d438a42eb177d9322212cde6ba8f367912a,Command:[],Args:[--prometheus-config-reloader=$(RELATED_IMAGE_PROMETHEUS_CONFIG_RELOADER) --prometheus-instance-selector=app.kubernetes.io/managed-by=observability-operator --alertmanager-instance-selector=app.kubernetes.io/managed-by=observability-operator --thanos-ruler-instance-selector=app.kubernetes.io/managed-by=observability-operator --watch-referenced-objects-in-all-namespaces=true --disable-unmanaged-prometheus-configuration=true],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:http,HostPort:0,ContainerPort:8080,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:GOGC,Value:30,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PROMETHEUS_CONFIG_RELOADER,Value:registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-prometheus-config-reloader-rhel9@sha256:9a2097bc5b2e02bc1703f64c452ce8fe4bc6775b732db930ff4770b76ae4653a,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.1,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{157286400 0} {} 150Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bmwsq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod obo-prometheus-operator-68bc856cb9-cq2gx_openshift-operators(e31d2d53-8992-45e3-98aa-24ea73236248): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 26 17:11:28 crc kubenswrapper[4856]: E0126 17:11:28.246600 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-cq2gx" podUID="e31d2d53-8992-45e3-98aa-24ea73236248" Jan 26 17:11:28 crc kubenswrapper[4856]: E0126 17:11:28.314628 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/obo-prometheus-rhel9-operator@sha256:e7e5f4c5e8ab0ba298ef0295a7137d438a42eb177d9322212cde6ba8f367912a\\\"\"" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-cq2gx" podUID="e31d2d53-8992-45e3-98aa-24ea73236248" Jan 26 17:11:36 crc kubenswrapper[4856]: E0126 17:11:36.180306 4856 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="registry.connect.redhat.com/elastic/eck-operator@sha256:28925fffef8f7c920b2510810cbcfc0f3dadab5f8a80b01fd5ae500e5c070105" Jan 26 17:11:36 crc kubenswrapper[4856]: E0126 17:11:36.181416 4856 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:registry.connect.redhat.com/elastic/eck-operator@sha256:28925fffef8f7c920b2510810cbcfc0f3dadab5f8a80b01fd5ae500e5c070105,Command:[],Args:[manager --config=/conf/eck.yaml --manage-webhook-certs=false --enable-webhook --ubi-only --distribution-channel=certified-operators],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https-webhook,HostPort:0,ContainerPort:9443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:NAMESPACES,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.annotations['olm.targetNamespaces'],},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.annotations['olm.operatorNamespace'],},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:OPERATOR_IMAGE,Value:registry.connect.redhat.com/elastic/eck-operator@sha256:28925fffef8f7c920b2510810cbcfc0f3dadab5f8a80b01fd5ae500e5c070105,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:elasticsearch-eck-operator-certified.v3.2.0,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{1 0} {} 1 DecimalSI},memory: {{1073741824 0} {} 1Gi BinarySI},},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{157286400 0} {} 150Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:apiservice-cert,ReadOnly:false,MountPath:/apiserver.local.config/certificates,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-z7wld,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000670000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod elastic-operator-bf765cf6c-gbst6_service-telemetry(9988655e-7b1f-443e-a102-68665719162a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 17:11:36 crc kubenswrapper[4856]: E0126 17:11:36.182633 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="service-telemetry/elastic-operator-bf765cf6c-gbst6" podUID="9988655e-7b1f-443e-a102-68665719162a" Jan 26 17:11:36 crc kubenswrapper[4856]: E0126 17:11:36.333656 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.connect.redhat.com/elastic/eck-operator@sha256:28925fffef8f7c920b2510810cbcfc0f3dadab5f8a80b01fd5ae500e5c070105\\\"\"" pod="service-telemetry/elastic-operator-bf765cf6c-gbst6" podUID="9988655e-7b1f-443e-a102-68665719162a" Jan 26 17:11:37 crc kubenswrapper[4856]: E0126 17:11:37.636497 4856 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:b5c8526d2ae660fe092dd8a7acf18ec4957d5c265890a222f55396fc2cdaeed8" Jan 26 17:11:37 crc kubenswrapper[4856]: E0126 17:11:37.637182 4856 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:perses-operator,Image:registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:b5c8526d2ae660fe092dd8a7acf18ec4957d5c265890a222f55396fc2cdaeed8,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.1,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{134217728 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:openshift-service-ca,ReadOnly:true,MountPath:/ca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tmjhr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000350000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod perses-operator-5bf474d74f-5bmfp_openshift-operators(bd7597f2-d44b-4e1b-ac60-b409985e3351): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 26 17:11:37 crc kubenswrapper[4856]: E0126 17:11:37.638432 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"perses-operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/perses-operator-5bf474d74f-5bmfp" podUID="bd7597f2-d44b-4e1b-ac60-b409985e3351" Jan 26 17:11:37 crc kubenswrapper[4856]: E0126 17:11:37.655579 4856 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:42ebc3571195d8c41fd01b8d08e98fe2cc12c1caabea251aecb4442d8eade4ea" Jan 26 17:11:37 crc kubenswrapper[4856]: E0126 17:11:37.655882 4856 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:prometheus-operator-admission-webhook,Image:registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:42ebc3571195d8c41fd01b8d08e98fe2cc12c1caabea251aecb4442d8eade4ea,Command:[],Args:[--web.enable-tls=true --web.cert-file=/tmp/k8s-webhook-server/serving-certs/tls.crt --web.key-file=/tmp/k8s-webhook-server/serving-certs/tls.key],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.3.1,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{209715200 0} {} BinarySI},},Requests:ResourceList{cpu: {{50 -3} {} 50m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:apiservice-cert,ReadOnly:false,MountPath:/apiserver.local.config/certificates,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod obo-prometheus-operator-admission-webhook-68d9fdc4dd-cf7wn_openshift-operators(7c88687f-1304-4709-b148-a196f0d0190d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 26 17:11:37 crc kubenswrapper[4856]: E0126 17:11:37.657656 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator-admission-webhook\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d9fdc4dd-cf7wn" podUID="7c88687f-1304-4709-b148-a196f0d0190d" Jan 26 17:11:38 crc kubenswrapper[4856]: E0126 17:11:38.432349 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"perses-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/perses-rhel9-operator@sha256:b5c8526d2ae660fe092dd8a7acf18ec4957d5c265890a222f55396fc2cdaeed8\\\"\"" pod="openshift-operators/perses-operator-5bf474d74f-5bmfp" podUID="bd7597f2-d44b-4e1b-ac60-b409985e3351" Jan 26 17:11:38 crc kubenswrapper[4856]: E0126 17:11:38.432570 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus-operator-admission-webhook\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/obo-prometheus-operator-admission-webhook-rhel9@sha256:42ebc3571195d8c41fd01b8d08e98fe2cc12c1caabea251aecb4442d8eade4ea\\\"\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d9fdc4dd-cf7wn" podUID="7c88687f-1304-4709-b148-a196f0d0190d" Jan 26 17:11:39 crc kubenswrapper[4856]: I0126 17:11:39.372554 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-wl4gw"] Jan 26 17:11:39 crc kubenswrapper[4856]: E0126 17:11:39.372937 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6521dc23-8f4e-452f-ae3e-167424fa3ed2" containerName="util" Jan 26 17:11:39 crc kubenswrapper[4856]: I0126 17:11:39.372959 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="6521dc23-8f4e-452f-ae3e-167424fa3ed2" containerName="util" Jan 26 17:11:39 crc kubenswrapper[4856]: E0126 17:11:39.372970 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6521dc23-8f4e-452f-ae3e-167424fa3ed2" containerName="extract" Jan 26 17:11:39 crc kubenswrapper[4856]: I0126 17:11:39.372977 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="6521dc23-8f4e-452f-ae3e-167424fa3ed2" containerName="extract" Jan 26 17:11:39 crc kubenswrapper[4856]: E0126 17:11:39.372988 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6521dc23-8f4e-452f-ae3e-167424fa3ed2" containerName="pull" Jan 26 17:11:39 crc kubenswrapper[4856]: I0126 17:11:39.372995 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="6521dc23-8f4e-452f-ae3e-167424fa3ed2" containerName="pull" Jan 26 17:11:39 crc kubenswrapper[4856]: I0126 17:11:39.373231 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="6521dc23-8f4e-452f-ae3e-167424fa3ed2" containerName="extract" Jan 26 17:11:39 crc kubenswrapper[4856]: I0126 17:11:39.375854 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-wl4gw" Jan 26 17:11:39 crc kubenswrapper[4856]: I0126 17:11:39.378370 4856 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-8gzcb" Jan 26 17:11:39 crc kubenswrapper[4856]: I0126 17:11:39.378739 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Jan 26 17:11:39 crc kubenswrapper[4856]: I0126 17:11:39.379645 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Jan 26 17:11:39 crc kubenswrapper[4856]: I0126 17:11:39.389674 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-wl4gw"] Jan 26 17:11:39 crc kubenswrapper[4856]: I0126 17:11:39.459563 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/a976d6d5-989b-49e8-bb9e-00c54dba078a-tmp\") pod \"cert-manager-operator-controller-manager-5446d6888b-wl4gw\" (UID: \"a976d6d5-989b-49e8-bb9e-00c54dba078a\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-wl4gw" Jan 26 17:11:39 crc kubenswrapper[4856]: I0126 17:11:39.459773 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sjx8\" (UniqueName: \"kubernetes.io/projected/a976d6d5-989b-49e8-bb9e-00c54dba078a-kube-api-access-6sjx8\") pod \"cert-manager-operator-controller-manager-5446d6888b-wl4gw\" (UID: \"a976d6d5-989b-49e8-bb9e-00c54dba078a\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-wl4gw" Jan 26 17:11:39 crc kubenswrapper[4856]: I0126 17:11:39.561360 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sjx8\" (UniqueName: \"kubernetes.io/projected/a976d6d5-989b-49e8-bb9e-00c54dba078a-kube-api-access-6sjx8\") pod \"cert-manager-operator-controller-manager-5446d6888b-wl4gw\" (UID: \"a976d6d5-989b-49e8-bb9e-00c54dba078a\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-wl4gw" Jan 26 17:11:39 crc kubenswrapper[4856]: I0126 17:11:39.561439 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/a976d6d5-989b-49e8-bb9e-00c54dba078a-tmp\") pod \"cert-manager-operator-controller-manager-5446d6888b-wl4gw\" (UID: \"a976d6d5-989b-49e8-bb9e-00c54dba078a\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-wl4gw" Jan 26 17:11:39 crc kubenswrapper[4856]: I0126 17:11:39.561922 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/a976d6d5-989b-49e8-bb9e-00c54dba078a-tmp\") pod \"cert-manager-operator-controller-manager-5446d6888b-wl4gw\" (UID: \"a976d6d5-989b-49e8-bb9e-00c54dba078a\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-wl4gw" Jan 26 17:11:39 crc kubenswrapper[4856]: I0126 17:11:39.585423 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sjx8\" (UniqueName: \"kubernetes.io/projected/a976d6d5-989b-49e8-bb9e-00c54dba078a-kube-api-access-6sjx8\") pod \"cert-manager-operator-controller-manager-5446d6888b-wl4gw\" (UID: \"a976d6d5-989b-49e8-bb9e-00c54dba078a\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-wl4gw" Jan 26 17:11:39 crc kubenswrapper[4856]: I0126 17:11:39.697309 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-wl4gw" Jan 26 17:11:43 crc kubenswrapper[4856]: E0126 17:11:43.249775 4856 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43" Jan 26 17:11:43 crc kubenswrapper[4856]: E0126 17:11:43.250020 4856 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:interconnect-operator,Image:registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43,Command:[qdr-operator],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:60000,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:WATCH_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:OPERATOR_NAME,Value:qdr-operator,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_QDROUTERD_IMAGE,Value:registry.redhat.io/amq7/amq-interconnect@sha256:31d87473fa684178a694f9ee331d3c80f2653f9533cb65c2a325752166a077e9,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:amq7-interconnect-operator.v1.10.20,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qhz7k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000670000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod interconnect-operator-5bb49f789d-k8bc4_service-telemetry(15358471-0b96-4792-aaaa-433823f9ed88): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 26 17:11:43 crc kubenswrapper[4856]: E0126 17:11:43.251226 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"interconnect-operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="service-telemetry/interconnect-operator-5bb49f789d-k8bc4" podUID="15358471-0b96-4792-aaaa-433823f9ed88" Jan 26 17:11:43 crc kubenswrapper[4856]: E0126 17:11:43.499227 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"interconnect-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43\\\"\"" pod="service-telemetry/interconnect-operator-5bb49f789d-k8bc4" podUID="15358471-0b96-4792-aaaa-433823f9ed88" Jan 26 17:11:43 crc kubenswrapper[4856]: I0126 17:11:43.544005 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-wl4gw"] Jan 26 17:11:43 crc kubenswrapper[4856]: W0126 17:11:43.550033 4856 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda976d6d5_989b_49e8_bb9e_00c54dba078a.slice/crio-c0badef7f49f1421504f385a874d4ddb46c0f67be5b4353c924de6ec8b397518 WatchSource:0}: Error finding container c0badef7f49f1421504f385a874d4ddb46c0f67be5b4353c924de6ec8b397518: Status 404 returned error can't find the container with id c0badef7f49f1421504f385a874d4ddb46c0f67be5b4353c924de6ec8b397518 Jan 26 17:11:44 crc kubenswrapper[4856]: I0126 17:11:44.504237 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d9fdc4dd-jbq25" event={"ID":"766f50ba-0751-4f25-a6db-3b7195e72f55","Type":"ContainerStarted","Data":"d4af2ae0fc9165956735abdaecbb6623aeb839ccd306f173315a8ca783224c65"} Jan 26 17:11:44 crc kubenswrapper[4856]: I0126 17:11:44.507441 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-cq2gx" event={"ID":"e31d2d53-8992-45e3-98aa-24ea73236248","Type":"ContainerStarted","Data":"8236a64fbd4a9d4a0bc8353f3388c89b02598fe759ed8fd5b7bf123bcc6a0723"} Jan 26 17:11:44 crc kubenswrapper[4856]: I0126 17:11:44.511003 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-fpn2h" event={"ID":"a4ae7646-2afb-4ada-b8a4-d20a69f87949","Type":"ContainerStarted","Data":"5760274324fbd1a6f9babf8c0cd31646fd5ec80c55fb710d1c0b47542e1ef0f1"} Jan 26 17:11:44 crc kubenswrapper[4856]: I0126 17:11:44.511236 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-59bdc8b94-fpn2h" Jan 26 17:11:44 crc kubenswrapper[4856]: I0126 17:11:44.513466 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-wl4gw" event={"ID":"a976d6d5-989b-49e8-bb9e-00c54dba078a","Type":"ContainerStarted","Data":"c0badef7f49f1421504f385a874d4ddb46c0f67be5b4353c924de6ec8b397518"} Jan 26 17:11:44 crc kubenswrapper[4856]: I0126 17:11:44.513668 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-59bdc8b94-fpn2h" Jan 26 17:11:44 crc kubenswrapper[4856]: I0126 17:11:44.537496 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d9fdc4dd-jbq25" podStartSLOduration=3.247349798 podStartE2EDuration="39.537461142s" podCreationTimestamp="2026-01-26 17:11:05 +0000 UTC" firstStartedPulling="2026-01-26 17:11:06.981105038 +0000 UTC m=+762.934359019" lastFinishedPulling="2026-01-26 17:11:43.271216382 +0000 UTC m=+799.224470363" observedRunningTime="2026-01-26 17:11:44.536454323 +0000 UTC m=+800.489708314" watchObservedRunningTime="2026-01-26 17:11:44.537461142 +0000 UTC m=+800.490715133" Jan 26 17:11:44 crc kubenswrapper[4856]: I0126 17:11:44.609974 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-59bdc8b94-fpn2h" podStartSLOduration=2.280301729 podStartE2EDuration="38.609955452s" podCreationTimestamp="2026-01-26 17:11:06 +0000 UTC" firstStartedPulling="2026-01-26 17:11:06.941563689 +0000 UTC m=+762.894817680" lastFinishedPulling="2026-01-26 17:11:43.271217422 +0000 UTC m=+799.224471403" observedRunningTime="2026-01-26 17:11:44.583604279 +0000 UTC m=+800.536858280" watchObservedRunningTime="2026-01-26 17:11:44.609955452 +0000 UTC m=+800.563209433" Jan 26 17:11:44 crc kubenswrapper[4856]: I0126 17:11:44.610261 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-cq2gx" podStartSLOduration=3.035226025 podStartE2EDuration="39.61025685s" podCreationTimestamp="2026-01-26 17:11:05 +0000 UTC" firstStartedPulling="2026-01-26 17:11:06.73584922 +0000 UTC m=+762.689103201" lastFinishedPulling="2026-01-26 17:11:43.310880045 +0000 UTC m=+799.264134026" observedRunningTime="2026-01-26 17:11:44.606007509 +0000 UTC m=+800.559261500" watchObservedRunningTime="2026-01-26 17:11:44.61025685 +0000 UTC m=+800.563510821" Jan 26 17:11:53 crc kubenswrapper[4856]: I0126 17:11:53.575093 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-wl4gw" event={"ID":"a976d6d5-989b-49e8-bb9e-00c54dba078a","Type":"ContainerStarted","Data":"21f00369be34e114c32e8e60d719fdedf1ab109abb693b3da0814d8133a1f758"} Jan 26 17:11:53 crc kubenswrapper[4856]: I0126 17:11:53.576677 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elastic-operator-bf765cf6c-gbst6" event={"ID":"9988655e-7b1f-443e-a102-68665719162a","Type":"ContainerStarted","Data":"74ec440aeea33ccacd27ac561070abd83a42057c3516d917c9525d4bdabe6435"} Jan 26 17:11:53 crc kubenswrapper[4856]: I0126 17:11:53.578644 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-5bmfp" event={"ID":"bd7597f2-d44b-4e1b-ac60-b409985e3351","Type":"ContainerStarted","Data":"66e10285f6d296dd2a1b78f58e5ac030da89973d1aac49cea6bf74e3ede6338e"} Jan 26 17:11:53 crc kubenswrapper[4856]: I0126 17:11:53.578940 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5bf474d74f-5bmfp" Jan 26 17:11:53 crc kubenswrapper[4856]: I0126 17:11:53.579596 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d9fdc4dd-cf7wn" event={"ID":"7c88687f-1304-4709-b148-a196f0d0190d","Type":"ContainerStarted","Data":"e28c16cd32a7cfcd39588bf2118b2218eb24d94557f0688fdf1303b55eb3bcfd"} Jan 26 17:11:53 crc kubenswrapper[4856]: I0126 17:11:53.668141 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-5446d6888b-wl4gw" podStartSLOduration=5.043201361 podStartE2EDuration="14.668108066s" podCreationTimestamp="2026-01-26 17:11:39 +0000 UTC" firstStartedPulling="2026-01-26 17:11:43.553938573 +0000 UTC m=+799.507192554" lastFinishedPulling="2026-01-26 17:11:53.178845288 +0000 UTC m=+809.132099259" observedRunningTime="2026-01-26 17:11:53.661981051 +0000 UTC m=+809.615235032" watchObservedRunningTime="2026-01-26 17:11:53.668108066 +0000 UTC m=+809.621362047" Jan 26 17:11:53 crc kubenswrapper[4856]: I0126 17:11:53.688239 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/elastic-operator-bf765cf6c-gbst6" podStartSLOduration=3.830375178 podStartE2EDuration="45.68822016s" podCreationTimestamp="2026-01-26 17:11:08 +0000 UTC" firstStartedPulling="2026-01-26 17:11:11.570094657 +0000 UTC m=+767.523348638" lastFinishedPulling="2026-01-26 17:11:53.427939639 +0000 UTC m=+809.381193620" observedRunningTime="2026-01-26 17:11:53.682356513 +0000 UTC m=+809.635610514" watchObservedRunningTime="2026-01-26 17:11:53.68822016 +0000 UTC m=+809.641474141" Jan 26 17:11:53 crc kubenswrapper[4856]: I0126 17:11:53.706868 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5bf474d74f-5bmfp" podStartSLOduration=1.3310320390000001 podStartE2EDuration="47.706849182s" podCreationTimestamp="2026-01-26 17:11:06 +0000 UTC" firstStartedPulling="2026-01-26 17:11:06.981620573 +0000 UTC m=+762.934874554" lastFinishedPulling="2026-01-26 17:11:53.357437726 +0000 UTC m=+809.310691697" observedRunningTime="2026-01-26 17:11:53.706012058 +0000 UTC m=+809.659266059" watchObservedRunningTime="2026-01-26 17:11:53.706849182 +0000 UTC m=+809.660103163" Jan 26 17:11:53 crc kubenswrapper[4856]: I0126 17:11:53.729923 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-68d9fdc4dd-cf7wn" podStartSLOduration=-9223371988.12487 podStartE2EDuration="48.7299051s" podCreationTimestamp="2026-01-26 17:11:05 +0000 UTC" firstStartedPulling="2026-01-26 17:11:06.865061054 +0000 UTC m=+762.818315035" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:11:53.726687488 +0000 UTC m=+809.679941479" watchObservedRunningTime="2026-01-26 17:11:53.7299051 +0000 UTC m=+809.683159081" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.692726 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.694589 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.707760 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"elasticsearch-es-unicast-hosts" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.707773 4856 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-remote-ca" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.708279 4856 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-default-es-transport-certs" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.708567 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"elasticsearch-es-scripts" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.709842 4856 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-default-es-config" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.710215 4856 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-xpack-file-realm" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.710447 4856 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-internal-users" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.710767 4856 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-dockercfg-5t26m" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.710864 4856 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-http-certs-internal" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.747722 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.784166 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.784222 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.784249 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.784290 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.784318 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.784335 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.784356 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.784387 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.784406 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.784421 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.784442 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.784466 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.784486 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.784547 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.784583 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.889249 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.889310 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.889344 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.889454 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.889517 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.889585 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.889611 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.889661 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.889694 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.889735 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.889782 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.889830 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.889866 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.889926 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.889973 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.895603 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.896631 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.901181 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.901230 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.901600 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.901935 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.903717 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.905326 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.905776 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.906130 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.906764 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.909190 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.909485 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.931127 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:55 crc kubenswrapper[4856]: I0126 17:11:55.936345 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/8cba7b0b-8fbc-4d94-a808-43c46e0defaa-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"8cba7b0b-8fbc-4d94-a808-43c46e0defaa\") " pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:56 crc kubenswrapper[4856]: I0126 17:11:56.143946 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:11:56 crc kubenswrapper[4856]: I0126 17:11:56.563747 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Jan 26 17:11:56 crc kubenswrapper[4856]: W0126 17:11:56.567231 4856 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8cba7b0b_8fbc_4d94_a808_43c46e0defaa.slice/crio-db3e40a1186e23ef0cae8e0d0a1825b9e17257d4c36b7abfe5dd7998d36153e6 WatchSource:0}: Error finding container db3e40a1186e23ef0cae8e0d0a1825b9e17257d4c36b7abfe5dd7998d36153e6: Status 404 returned error can't find the container with id db3e40a1186e23ef0cae8e0d0a1825b9e17257d4c36b7abfe5dd7998d36153e6 Jan 26 17:11:56 crc kubenswrapper[4856]: I0126 17:11:56.738597 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"8cba7b0b-8fbc-4d94-a808-43c46e0defaa","Type":"ContainerStarted","Data":"db3e40a1186e23ef0cae8e0d0a1825b9e17257d4c36b7abfe5dd7998d36153e6"} Jan 26 17:11:56 crc kubenswrapper[4856]: I0126 17:11:56.952681 4856 patch_prober.go:28] interesting pod/machine-config-daemon-xm9cq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:11:56 crc kubenswrapper[4856]: I0126 17:11:56.952756 4856 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:11:57 crc kubenswrapper[4856]: I0126 17:11:57.164025 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-www8b"] Jan 26 17:11:57 crc kubenswrapper[4856]: I0126 17:11:57.165073 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-www8b" Jan 26 17:11:57 crc kubenswrapper[4856]: I0126 17:11:57.168613 4856 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-r9d9k" Jan 26 17:11:57 crc kubenswrapper[4856]: I0126 17:11:57.169056 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Jan 26 17:11:57 crc kubenswrapper[4856]: I0126 17:11:57.169339 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Jan 26 17:11:57 crc kubenswrapper[4856]: I0126 17:11:57.176506 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-www8b"] Jan 26 17:11:57 crc kubenswrapper[4856]: I0126 17:11:57.289378 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e9288910-baf7-4cc4-b313-c87b80bfdd3e-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-www8b\" (UID: \"e9288910-baf7-4cc4-b313-c87b80bfdd3e\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-www8b" Jan 26 17:11:57 crc kubenswrapper[4856]: I0126 17:11:57.289539 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pw4xs\" (UniqueName: \"kubernetes.io/projected/e9288910-baf7-4cc4-b313-c87b80bfdd3e-kube-api-access-pw4xs\") pod \"cert-manager-webhook-f4fb5df64-www8b\" (UID: \"e9288910-baf7-4cc4-b313-c87b80bfdd3e\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-www8b" Jan 26 17:11:57 crc kubenswrapper[4856]: I0126 17:11:57.390974 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e9288910-baf7-4cc4-b313-c87b80bfdd3e-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-www8b\" (UID: \"e9288910-baf7-4cc4-b313-c87b80bfdd3e\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-www8b" Jan 26 17:11:57 crc kubenswrapper[4856]: I0126 17:11:57.391059 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pw4xs\" (UniqueName: \"kubernetes.io/projected/e9288910-baf7-4cc4-b313-c87b80bfdd3e-kube-api-access-pw4xs\") pod \"cert-manager-webhook-f4fb5df64-www8b\" (UID: \"e9288910-baf7-4cc4-b313-c87b80bfdd3e\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-www8b" Jan 26 17:11:57 crc kubenswrapper[4856]: I0126 17:11:57.414649 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e9288910-baf7-4cc4-b313-c87b80bfdd3e-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-www8b\" (UID: \"e9288910-baf7-4cc4-b313-c87b80bfdd3e\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-www8b" Jan 26 17:11:57 crc kubenswrapper[4856]: I0126 17:11:57.424416 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pw4xs\" (UniqueName: \"kubernetes.io/projected/e9288910-baf7-4cc4-b313-c87b80bfdd3e-kube-api-access-pw4xs\") pod \"cert-manager-webhook-f4fb5df64-www8b\" (UID: \"e9288910-baf7-4cc4-b313-c87b80bfdd3e\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-www8b" Jan 26 17:11:57 crc kubenswrapper[4856]: I0126 17:11:57.499890 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-www8b" Jan 26 17:11:57 crc kubenswrapper[4856]: I0126 17:11:57.949963 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-www8b"] Jan 26 17:11:57 crc kubenswrapper[4856]: W0126 17:11:57.962563 4856 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9288910_baf7_4cc4_b313_c87b80bfdd3e.slice/crio-ddd5416f5cb3a66e434992028c974aa69e0cdcf6a9d28626bf0068580e89a28b WatchSource:0}: Error finding container ddd5416f5cb3a66e434992028c974aa69e0cdcf6a9d28626bf0068580e89a28b: Status 404 returned error can't find the container with id ddd5416f5cb3a66e434992028c974aa69e0cdcf6a9d28626bf0068580e89a28b Jan 26 17:11:58 crc kubenswrapper[4856]: I0126 17:11:58.620597 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-rm9wd"] Jan 26 17:11:58 crc kubenswrapper[4856]: I0126 17:11:58.626694 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-rm9wd" Jan 26 17:11:58 crc kubenswrapper[4856]: I0126 17:11:58.629182 4856 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-4m8h6" Jan 26 17:11:58 crc kubenswrapper[4856]: I0126 17:11:58.630177 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-rm9wd"] Jan 26 17:11:58 crc kubenswrapper[4856]: I0126 17:11:58.809127 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-www8b" event={"ID":"e9288910-baf7-4cc4-b313-c87b80bfdd3e","Type":"ContainerStarted","Data":"ddd5416f5cb3a66e434992028c974aa69e0cdcf6a9d28626bf0068580e89a28b"} Jan 26 17:11:58 crc kubenswrapper[4856]: I0126 17:11:58.813507 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kflgh\" (UniqueName: \"kubernetes.io/projected/24a9d780-2b57-49d2-9cb9-eac2456ed86d-kube-api-access-kflgh\") pod \"cert-manager-cainjector-855d9ccff4-rm9wd\" (UID: \"24a9d780-2b57-49d2-9cb9-eac2456ed86d\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-rm9wd" Jan 26 17:11:58 crc kubenswrapper[4856]: I0126 17:11:58.813650 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/24a9d780-2b57-49d2-9cb9-eac2456ed86d-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-rm9wd\" (UID: \"24a9d780-2b57-49d2-9cb9-eac2456ed86d\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-rm9wd" Jan 26 17:11:58 crc kubenswrapper[4856]: I0126 17:11:58.915704 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/24a9d780-2b57-49d2-9cb9-eac2456ed86d-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-rm9wd\" (UID: \"24a9d780-2b57-49d2-9cb9-eac2456ed86d\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-rm9wd" Jan 26 17:11:58 crc kubenswrapper[4856]: I0126 17:11:58.915798 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kflgh\" (UniqueName: \"kubernetes.io/projected/24a9d780-2b57-49d2-9cb9-eac2456ed86d-kube-api-access-kflgh\") pod \"cert-manager-cainjector-855d9ccff4-rm9wd\" (UID: \"24a9d780-2b57-49d2-9cb9-eac2456ed86d\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-rm9wd" Jan 26 17:11:58 crc kubenswrapper[4856]: I0126 17:11:58.946302 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kflgh\" (UniqueName: \"kubernetes.io/projected/24a9d780-2b57-49d2-9cb9-eac2456ed86d-kube-api-access-kflgh\") pod \"cert-manager-cainjector-855d9ccff4-rm9wd\" (UID: \"24a9d780-2b57-49d2-9cb9-eac2456ed86d\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-rm9wd" Jan 26 17:11:58 crc kubenswrapper[4856]: I0126 17:11:58.987869 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/24a9d780-2b57-49d2-9cb9-eac2456ed86d-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-rm9wd\" (UID: \"24a9d780-2b57-49d2-9cb9-eac2456ed86d\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-rm9wd" Jan 26 17:11:59 crc kubenswrapper[4856]: I0126 17:11:59.244154 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-rm9wd" Jan 26 17:11:59 crc kubenswrapper[4856]: I0126 17:11:59.838165 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/interconnect-operator-5bb49f789d-k8bc4" event={"ID":"15358471-0b96-4792-aaaa-433823f9ed88","Type":"ContainerStarted","Data":"97dcbb99650087b040f160ca42a2bd8e63017c9cb21286cf505b461eae5754e5"} Jan 26 17:11:59 crc kubenswrapper[4856]: I0126 17:11:59.872921 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/interconnect-operator-5bb49f789d-k8bc4" podStartSLOduration=1.734664881 podStartE2EDuration="47.872899048s" podCreationTimestamp="2026-01-26 17:11:12 +0000 UTC" firstStartedPulling="2026-01-26 17:11:13.092200689 +0000 UTC m=+769.045454670" lastFinishedPulling="2026-01-26 17:11:59.230434856 +0000 UTC m=+815.183688837" observedRunningTime="2026-01-26 17:11:59.87017467 +0000 UTC m=+815.823428661" watchObservedRunningTime="2026-01-26 17:11:59.872899048 +0000 UTC m=+815.826153039" Jan 26 17:12:00 crc kubenswrapper[4856]: I0126 17:12:00.033493 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-rm9wd"] Jan 26 17:12:00 crc kubenswrapper[4856]: I0126 17:12:00.846772 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-rm9wd" event={"ID":"24a9d780-2b57-49d2-9cb9-eac2456ed86d","Type":"ContainerStarted","Data":"e6ff1bea74c7a653ac5b07ac4eaf01f57fa00dc9d26acd07d07c7c3048e0d12f"} Jan 26 17:12:06 crc kubenswrapper[4856]: I0126 17:12:06.695273 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5bf474d74f-5bmfp" Jan 26 17:12:16 crc kubenswrapper[4856]: I0126 17:12:16.409793 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-86cb77c54b-xcqr4"] Jan 26 17:12:16 crc kubenswrapper[4856]: I0126 17:12:16.411833 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-xcqr4" Jan 26 17:12:16 crc kubenswrapper[4856]: I0126 17:12:16.415603 4856 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-gxmx7" Jan 26 17:12:16 crc kubenswrapper[4856]: I0126 17:12:16.493615 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-xcqr4"] Jan 26 17:12:16 crc kubenswrapper[4856]: I0126 17:12:16.513505 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92jf9\" (UniqueName: \"kubernetes.io/projected/3dc10d6b-aa48-4c7d-afab-45fa62298819-kube-api-access-92jf9\") pod \"cert-manager-86cb77c54b-xcqr4\" (UID: \"3dc10d6b-aa48-4c7d-afab-45fa62298819\") " pod="cert-manager/cert-manager-86cb77c54b-xcqr4" Jan 26 17:12:16 crc kubenswrapper[4856]: I0126 17:12:16.513597 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3dc10d6b-aa48-4c7d-afab-45fa62298819-bound-sa-token\") pod \"cert-manager-86cb77c54b-xcqr4\" (UID: \"3dc10d6b-aa48-4c7d-afab-45fa62298819\") " pod="cert-manager/cert-manager-86cb77c54b-xcqr4" Jan 26 17:12:16 crc kubenswrapper[4856]: I0126 17:12:16.615642 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92jf9\" (UniqueName: \"kubernetes.io/projected/3dc10d6b-aa48-4c7d-afab-45fa62298819-kube-api-access-92jf9\") pod \"cert-manager-86cb77c54b-xcqr4\" (UID: \"3dc10d6b-aa48-4c7d-afab-45fa62298819\") " pod="cert-manager/cert-manager-86cb77c54b-xcqr4" Jan 26 17:12:16 crc kubenswrapper[4856]: I0126 17:12:16.616015 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3dc10d6b-aa48-4c7d-afab-45fa62298819-bound-sa-token\") pod \"cert-manager-86cb77c54b-xcqr4\" (UID: \"3dc10d6b-aa48-4c7d-afab-45fa62298819\") " pod="cert-manager/cert-manager-86cb77c54b-xcqr4" Jan 26 17:12:16 crc kubenswrapper[4856]: I0126 17:12:16.635516 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3dc10d6b-aa48-4c7d-afab-45fa62298819-bound-sa-token\") pod \"cert-manager-86cb77c54b-xcqr4\" (UID: \"3dc10d6b-aa48-4c7d-afab-45fa62298819\") " pod="cert-manager/cert-manager-86cb77c54b-xcqr4" Jan 26 17:12:16 crc kubenswrapper[4856]: I0126 17:12:16.635585 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92jf9\" (UniqueName: \"kubernetes.io/projected/3dc10d6b-aa48-4c7d-afab-45fa62298819-kube-api-access-92jf9\") pod \"cert-manager-86cb77c54b-xcqr4\" (UID: \"3dc10d6b-aa48-4c7d-afab-45fa62298819\") " pod="cert-manager/cert-manager-86cb77c54b-xcqr4" Jan 26 17:12:16 crc kubenswrapper[4856]: I0126 17:12:16.744853 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-xcqr4" Jan 26 17:12:17 crc kubenswrapper[4856]: E0126 17:12:17.436279 4856 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:29a0fa1c2f2a6cee62a0468a3883d16d491b4af29130dad6e3e2bb2948f274df" Jan 26 17:12:17 crc kubenswrapper[4856]: E0126 17:12:17.436606 4856 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cert-manager-webhook,Image:registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:29a0fa1c2f2a6cee62a0468a3883d16d491b4af29130dad6e3e2bb2948f274df,Command:[/app/cmd/webhook/webhook],Args:[--dynamic-serving-ca-secret-name=cert-manager-webhook-ca --dynamic-serving-ca-secret-namespace=$(POD_NAMESPACE) --dynamic-serving-dns-names=cert-manager-webhook,cert-manager-webhook.$(POD_NAMESPACE),cert-manager-webhook.$(POD_NAMESPACE).svc --secure-port=10250 --v=2],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:10250,Protocol:TCP,HostIP:,},ContainerPort{Name:healthcheck,HostPort:0,ContainerPort:6080,Protocol:TCP,HostIP:,},ContainerPort{Name:http-metrics,HostPort:0,ContainerPort:9402,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:POD_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:bound-sa-token,ReadOnly:true,MountPath:/var/run/secrets/openshift/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pw4xs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/livez,Port:{1 0 healthcheck},Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:60,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{1 0 healthcheck},Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000690000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cert-manager-webhook-f4fb5df64-www8b_cert-manager(e9288910-baf7-4cc4-b313-c87b80bfdd3e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 26 17:12:17 crc kubenswrapper[4856]: E0126 17:12:17.437794 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cert-manager-webhook\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="cert-manager/cert-manager-webhook-f4fb5df64-www8b" podUID="e9288910-baf7-4cc4-b313-c87b80bfdd3e" Jan 26 17:12:18 crc kubenswrapper[4856]: E0126 17:12:18.349107 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cert-manager-webhook\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cert-manager/jetstack-cert-manager-rhel9@sha256:29a0fa1c2f2a6cee62a0468a3883d16d491b4af29130dad6e3e2bb2948f274df\\\"\"" pod="cert-manager/cert-manager-webhook-f4fb5df64-www8b" podUID="e9288910-baf7-4cc4-b313-c87b80bfdd3e" Jan 26 17:12:26 crc kubenswrapper[4856]: E0126 17:12:26.720227 4856 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="registry.connect.redhat.com/elastic/elasticsearch:7.17.20" Jan 26 17:12:26 crc kubenswrapper[4856]: E0126 17:12:26.721032 4856 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:elastic-internal-init-filesystem,Image:registry.connect.redhat.com/elastic/elasticsearch:7.17.20,Command:[bash -c /mnt/elastic-internal/scripts/prepare-fs.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:HEADLESS_SERVICE_NAME,Value:elasticsearch-es-default,ValueFrom:nil,},EnvVar{Name:PROBE_PASSWORD_PATH,Value:/mnt/elastic-internal/pod-mounted-users/elastic-internal-probe,ValueFrom:nil,},EnvVar{Name:PROBE_USERNAME,Value:elastic-internal-probe,ValueFrom:nil,},EnvVar{Name:READINESS_PROBE_PROTOCOL,Value:https,ValueFrom:nil,},EnvVar{Name:NSS_SDB_USE_CACHE,Value:no,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:downward-api,ReadOnly:true,MountPath:/mnt/elastic-internal/downward-api,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-elasticsearch-bin-local,ReadOnly:false,MountPath:/mnt/elastic-internal/elasticsearch-bin-local,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-elasticsearch-config,ReadOnly:true,MountPath:/mnt/elastic-internal/elasticsearch-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-elasticsearch-config-local,ReadOnly:false,MountPath:/mnt/elastic-internal/elasticsearch-config-local,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-elasticsearch-plugins-local,ReadOnly:false,MountPath:/mnt/elastic-internal/elasticsearch-plugins-local,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-http-certificates,ReadOnly:true,MountPath:/usr/share/elasticsearch/config/http-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-probe-user,ReadOnly:true,MountPath:/mnt/elastic-internal/pod-mounted-users,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-remote-certificate-authorities,ReadOnly:true,MountPath:/usr/share/elasticsearch/config/transport-remote-certs/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-scripts,ReadOnly:true,MountPath:/mnt/elastic-internal/scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-transport-certificates,ReadOnly:true,MountPath:/mnt/elastic-internal/transport-certificates,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-unicast-hosts,ReadOnly:true,MountPath:/mnt/elastic-internal/unicast-hosts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-xpack-file-realm,ReadOnly:true,MountPath:/mnt/elastic-internal/xpack-file-realm,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elasticsearch-data,ReadOnly:false,MountPath:/usr/share/elasticsearch/data,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elasticsearch-logs,ReadOnly:false,MountPath:/usr/share/elasticsearch/logs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tmp-volume,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*1000670000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod elasticsearch-es-default-0_service-telemetry(8cba7b0b-8fbc-4d94-a808-43c46e0defaa): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 17:12:26 crc kubenswrapper[4856]: E0126 17:12:26.722765 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"elastic-internal-init-filesystem\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="service-telemetry/elasticsearch-es-default-0" podUID="8cba7b0b-8fbc-4d94-a808-43c46e0defaa" Jan 26 17:12:26 crc kubenswrapper[4856]: I0126 17:12:26.767940 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-xcqr4"] Jan 26 17:12:27 crc kubenswrapper[4856]: I0126 17:12:26.939049 4856 patch_prober.go:28] interesting pod/machine-config-daemon-xm9cq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:12:27 crc kubenswrapper[4856]: I0126 17:12:26.939143 4856 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:12:27 crc kubenswrapper[4856]: I0126 17:12:27.591846 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-xcqr4" event={"ID":"3dc10d6b-aa48-4c7d-afab-45fa62298819","Type":"ContainerStarted","Data":"01718ae2682246101d0243d503913e773a0a7428a8847cd99c245f99820c6b31"} Jan 26 17:12:27 crc kubenswrapper[4856]: I0126 17:12:27.592136 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-xcqr4" event={"ID":"3dc10d6b-aa48-4c7d-afab-45fa62298819","Type":"ContainerStarted","Data":"b3ceac6859cf4cde100510b0215cca6038a853f9300b66b4346aa458836b853a"} Jan 26 17:12:27 crc kubenswrapper[4856]: I0126 17:12:27.594579 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-rm9wd" event={"ID":"24a9d780-2b57-49d2-9cb9-eac2456ed86d","Type":"ContainerStarted","Data":"c1544f4bd1c240b53131c6d9efe66c2aa6b12f5da34a643f11683134b5ae43a4"} Jan 26 17:12:27 crc kubenswrapper[4856]: E0126 17:12:27.595205 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"elastic-internal-init-filesystem\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.connect.redhat.com/elastic/elasticsearch:7.17.20\\\"\"" pod="service-telemetry/elasticsearch-es-default-0" podUID="8cba7b0b-8fbc-4d94-a808-43c46e0defaa" Jan 26 17:12:27 crc kubenswrapper[4856]: I0126 17:12:27.673916 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-86cb77c54b-xcqr4" podStartSLOduration=11.67387249 podStartE2EDuration="11.67387249s" podCreationTimestamp="2026-01-26 17:12:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:12:27.611896564 +0000 UTC m=+843.565150555" watchObservedRunningTime="2026-01-26 17:12:27.67387249 +0000 UTC m=+843.627126471" Jan 26 17:12:27 crc kubenswrapper[4856]: I0126 17:12:27.678910 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-855d9ccff4-rm9wd" podStartSLOduration=3.383466221 podStartE2EDuration="29.678889547s" podCreationTimestamp="2026-01-26 17:11:58 +0000 UTC" firstStartedPulling="2026-01-26 17:12:00.052535527 +0000 UTC m=+816.005789508" lastFinishedPulling="2026-01-26 17:12:26.347958853 +0000 UTC m=+842.301212834" observedRunningTime="2026-01-26 17:12:27.678540267 +0000 UTC m=+843.631794268" watchObservedRunningTime="2026-01-26 17:12:27.678889547 +0000 UTC m=+843.632143548" Jan 26 17:12:27 crc kubenswrapper[4856]: I0126 17:12:27.746719 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Jan 26 17:12:27 crc kubenswrapper[4856]: I0126 17:12:27.782258 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Jan 26 17:12:28 crc kubenswrapper[4856]: E0126 17:12:28.606401 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"elastic-internal-init-filesystem\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.connect.redhat.com/elastic/elasticsearch:7.17.20\\\"\"" pod="service-telemetry/elasticsearch-es-default-0" podUID="8cba7b0b-8fbc-4d94-a808-43c46e0defaa" Jan 26 17:12:29 crc kubenswrapper[4856]: E0126 17:12:29.610130 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"elastic-internal-init-filesystem\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.connect.redhat.com/elastic/elasticsearch:7.17.20\\\"\"" pod="service-telemetry/elasticsearch-es-default-0" podUID="8cba7b0b-8fbc-4d94-a808-43c46e0defaa" Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.307855 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.309074 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.311320 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-1-sys-config" Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.311320 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-1-ca" Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.311321 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-1-global-ca" Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.312296 4856 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-8h4xs" Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.439139 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-buildcachedir\") pod \"service-telemetry-operator-1-build\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.439191 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-container-storage-root\") pod \"service-telemetry-operator-1-build\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.439228 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-node-pullsecrets\") pod \"service-telemetry-operator-1-build\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.440682 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-build-blob-cache\") pod \"service-telemetry-operator-1-build\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.440713 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-build-proxy-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.440752 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsm2b\" (UniqueName: \"kubernetes.io/projected/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-kube-api-access-gsm2b\") pod \"service-telemetry-operator-1-build\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.440779 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-container-storage-run\") pod \"service-telemetry-operator-1-build\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.440796 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-buildworkdir\") pod \"service-telemetry-operator-1-build\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.440810 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-builder-dockercfg-8h4xs-push\") pod \"service-telemetry-operator-1-build\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.440842 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-build-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.440864 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-builder-dockercfg-8h4xs-pull\") pod \"service-telemetry-operator-1-build\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.440884 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-build-system-configs\") pod \"service-telemetry-operator-1-build\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.459968 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.541654 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-builder-dockercfg-8h4xs-pull\") pod \"service-telemetry-operator-1-build\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.541706 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-build-system-configs\") pod \"service-telemetry-operator-1-build\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.541733 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-buildcachedir\") pod \"service-telemetry-operator-1-build\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.541757 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-container-storage-root\") pod \"service-telemetry-operator-1-build\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.541791 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-node-pullsecrets\") pod \"service-telemetry-operator-1-build\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.541865 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-build-blob-cache\") pod \"service-telemetry-operator-1-build\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.541892 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-build-proxy-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.541930 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsm2b\" (UniqueName: \"kubernetes.io/projected/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-kube-api-access-gsm2b\") pod \"service-telemetry-operator-1-build\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.541975 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-container-storage-run\") pod \"service-telemetry-operator-1-build\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.541998 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-buildworkdir\") pod \"service-telemetry-operator-1-build\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.542016 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-builder-dockercfg-8h4xs-push\") pod \"service-telemetry-operator-1-build\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.542050 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-build-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.542716 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-build-blob-cache\") pod \"service-telemetry-operator-1-build\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.542812 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-buildworkdir\") pod \"service-telemetry-operator-1-build\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.542904 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-build-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.542928 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-build-proxy-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.543042 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-buildcachedir\") pod \"service-telemetry-operator-1-build\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.543243 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-build-system-configs\") pod \"service-telemetry-operator-1-build\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.543249 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-node-pullsecrets\") pod \"service-telemetry-operator-1-build\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.543428 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-container-storage-root\") pod \"service-telemetry-operator-1-build\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.543794 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-container-storage-run\") pod \"service-telemetry-operator-1-build\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.548217 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-builder-dockercfg-8h4xs-pull\") pod \"service-telemetry-operator-1-build\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.548290 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-builder-dockercfg-8h4xs-push\") pod \"service-telemetry-operator-1-build\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.563043 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsm2b\" (UniqueName: \"kubernetes.io/projected/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-kube-api-access-gsm2b\") pod \"service-telemetry-operator-1-build\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " pod="service-telemetry/service-telemetry-operator-1-build" Jan 26 17:12:31 crc kubenswrapper[4856]: I0126 17:12:31.744337 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Jan 26 17:12:32 crc kubenswrapper[4856]: I0126 17:12:32.240443 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Jan 26 17:12:32 crc kubenswrapper[4856]: I0126 17:12:32.629160 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"c90021ae-76d4-45c7-a3b8-04e3b9bbddce","Type":"ContainerStarted","Data":"dd686a65c0eba5b54c5e815a6d4137b24d38f89f6901d1d1e3907daf3f0a3f46"} Jan 26 17:12:33 crc kubenswrapper[4856]: I0126 17:12:33.637315 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-www8b" event={"ID":"e9288910-baf7-4cc4-b313-c87b80bfdd3e","Type":"ContainerStarted","Data":"772a0c9d1eb894fb3a5347dc5ee0598386b020c26e7d289fcc87645f6477eef8"} Jan 26 17:12:33 crc kubenswrapper[4856]: I0126 17:12:33.638781 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-f4fb5df64-www8b" Jan 26 17:12:33 crc kubenswrapper[4856]: I0126 17:12:33.668055 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-f4fb5df64-www8b" podStartSLOduration=-9223372000.186808 podStartE2EDuration="36.667968226s" podCreationTimestamp="2026-01-26 17:11:57 +0000 UTC" firstStartedPulling="2026-01-26 17:11:57.965107341 +0000 UTC m=+813.918361322" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:12:33.662622889 +0000 UTC m=+849.615876960" watchObservedRunningTime="2026-01-26 17:12:33.667968226 +0000 UTC m=+849.621222207" Jan 26 17:12:39 crc kubenswrapper[4856]: I0126 17:12:39.677666 4856 generic.go:334] "Generic (PLEG): container finished" podID="c90021ae-76d4-45c7-a3b8-04e3b9bbddce" containerID="28aff785d4bd8d892e6c32fa63d73c371a6f2a38a3d0f6ac418a977bacaecdd0" exitCode=0 Jan 26 17:12:39 crc kubenswrapper[4856]: I0126 17:12:39.677782 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"c90021ae-76d4-45c7-a3b8-04e3b9bbddce","Type":"ContainerDied","Data":"28aff785d4bd8d892e6c32fa63d73c371a6f2a38a3d0f6ac418a977bacaecdd0"} Jan 26 17:12:40 crc kubenswrapper[4856]: I0126 17:12:40.687690 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"c90021ae-76d4-45c7-a3b8-04e3b9bbddce","Type":"ContainerStarted","Data":"2800b2e70dad7de0c38ad3d8f25f147fcc58b2b450a62480a1e7bc1db9736970"} Jan 26 17:12:40 crc kubenswrapper[4856]: I0126 17:12:40.714286 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/service-telemetry-operator-1-build" podStartSLOduration=3.3986747250000002 podStartE2EDuration="9.714259875s" podCreationTimestamp="2026-01-26 17:12:31 +0000 UTC" firstStartedPulling="2026-01-26 17:12:32.249817081 +0000 UTC m=+848.203071062" lastFinishedPulling="2026-01-26 17:12:38.565402221 +0000 UTC m=+854.518656212" observedRunningTime="2026-01-26 17:12:40.707214282 +0000 UTC m=+856.660468273" watchObservedRunningTime="2026-01-26 17:12:40.714259875 +0000 UTC m=+856.667513856" Jan 26 17:12:41 crc kubenswrapper[4856]: I0126 17:12:41.457551 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Jan 26 17:12:42 crc kubenswrapper[4856]: I0126 17:12:42.503304 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-f4fb5df64-www8b" Jan 26 17:12:42 crc kubenswrapper[4856]: I0126 17:12:42.701813 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/service-telemetry-operator-1-build" podUID="c90021ae-76d4-45c7-a3b8-04e3b9bbddce" containerName="docker-build" containerID="cri-o://2800b2e70dad7de0c38ad3d8f25f147fcc58b2b450a62480a1e7bc1db9736970" gracePeriod=30 Jan 26 17:12:43 crc kubenswrapper[4856]: I0126 17:12:43.069992 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Jan 26 17:12:43 crc kubenswrapper[4856]: I0126 17:12:43.071998 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Jan 26 17:12:43 crc kubenswrapper[4856]: I0126 17:12:43.074264 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-2-global-ca" Jan 26 17:12:43 crc kubenswrapper[4856]: I0126 17:12:43.074319 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-2-ca" Jan 26 17:12:43 crc kubenswrapper[4856]: I0126 17:12:43.074276 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-2-sys-config" Jan 26 17:12:43 crc kubenswrapper[4856]: I0126 17:12:43.100655 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Jan 26 17:12:43 crc kubenswrapper[4856]: I0126 17:12:43.104290 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-build-blob-cache\") pod \"service-telemetry-operator-2-build\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 26 17:12:43 crc kubenswrapper[4856]: I0126 17:12:43.104348 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-node-pullsecrets\") pod \"service-telemetry-operator-2-build\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 26 17:12:43 crc kubenswrapper[4856]: I0126 17:12:43.104371 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-build-system-configs\") pod \"service-telemetry-operator-2-build\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 26 17:12:43 crc kubenswrapper[4856]: I0126 17:12:43.104391 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-container-storage-run\") pod \"service-telemetry-operator-2-build\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 26 17:12:43 crc kubenswrapper[4856]: I0126 17:12:43.104411 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-buildcachedir\") pod \"service-telemetry-operator-2-build\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 26 17:12:43 crc kubenswrapper[4856]: I0126 17:12:43.104442 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-builder-dockercfg-8h4xs-pull\") pod \"service-telemetry-operator-2-build\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 26 17:12:43 crc kubenswrapper[4856]: I0126 17:12:43.104471 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-buildworkdir\") pod \"service-telemetry-operator-2-build\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 26 17:12:43 crc kubenswrapper[4856]: I0126 17:12:43.104549 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-builder-dockercfg-8h4xs-push\") pod \"service-telemetry-operator-2-build\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 26 17:12:43 crc kubenswrapper[4856]: I0126 17:12:43.104573 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-container-storage-root\") pod \"service-telemetry-operator-2-build\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 26 17:12:43 crc kubenswrapper[4856]: I0126 17:12:43.104602 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spxgc\" (UniqueName: \"kubernetes.io/projected/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-kube-api-access-spxgc\") pod \"service-telemetry-operator-2-build\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 26 17:12:43 crc kubenswrapper[4856]: I0126 17:12:43.104646 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-build-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 26 17:12:43 crc kubenswrapper[4856]: I0126 17:12:43.104698 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-build-proxy-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 26 17:12:43 crc kubenswrapper[4856]: I0126 17:12:43.205752 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-build-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 26 17:12:43 crc kubenswrapper[4856]: I0126 17:12:43.205843 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-build-proxy-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 26 17:12:43 crc kubenswrapper[4856]: I0126 17:12:43.205908 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-build-blob-cache\") pod \"service-telemetry-operator-2-build\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 26 17:12:43 crc kubenswrapper[4856]: I0126 17:12:43.205950 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-node-pullsecrets\") pod \"service-telemetry-operator-2-build\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 26 17:12:43 crc kubenswrapper[4856]: I0126 17:12:43.205984 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-build-system-configs\") pod \"service-telemetry-operator-2-build\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 26 17:12:43 crc kubenswrapper[4856]: I0126 17:12:43.206012 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-container-storage-run\") pod \"service-telemetry-operator-2-build\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 26 17:12:43 crc kubenswrapper[4856]: I0126 17:12:43.206043 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-buildcachedir\") pod \"service-telemetry-operator-2-build\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 26 17:12:43 crc kubenswrapper[4856]: I0126 17:12:43.206074 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-builder-dockercfg-8h4xs-pull\") pod \"service-telemetry-operator-2-build\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 26 17:12:43 crc kubenswrapper[4856]: I0126 17:12:43.206132 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-buildworkdir\") pod \"service-telemetry-operator-2-build\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 26 17:12:43 crc kubenswrapper[4856]: I0126 17:12:43.206210 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-builder-dockercfg-8h4xs-push\") pod \"service-telemetry-operator-2-build\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 26 17:12:43 crc kubenswrapper[4856]: I0126 17:12:43.206271 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-container-storage-root\") pod \"service-telemetry-operator-2-build\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 26 17:12:43 crc kubenswrapper[4856]: I0126 17:12:43.206312 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spxgc\" (UniqueName: \"kubernetes.io/projected/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-kube-api-access-spxgc\") pod \"service-telemetry-operator-2-build\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 26 17:12:43 crc kubenswrapper[4856]: I0126 17:12:43.206600 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-buildcachedir\") pod \"service-telemetry-operator-2-build\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 26 17:12:43 crc kubenswrapper[4856]: I0126 17:12:43.206653 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-node-pullsecrets\") pod \"service-telemetry-operator-2-build\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 26 17:12:43 crc kubenswrapper[4856]: I0126 17:12:43.206960 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-build-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 26 17:12:43 crc kubenswrapper[4856]: I0126 17:12:43.248336 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-build-proxy-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 26 17:12:43 crc kubenswrapper[4856]: I0126 17:12:43.248389 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-container-storage-run\") pod \"service-telemetry-operator-2-build\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 26 17:12:43 crc kubenswrapper[4856]: I0126 17:12:43.249778 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-build-blob-cache\") pod \"service-telemetry-operator-2-build\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 26 17:12:43 crc kubenswrapper[4856]: I0126 17:12:43.249871 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-container-storage-root\") pod \"service-telemetry-operator-2-build\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 26 17:12:43 crc kubenswrapper[4856]: I0126 17:12:43.250263 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-buildworkdir\") pod \"service-telemetry-operator-2-build\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 26 17:12:43 crc kubenswrapper[4856]: I0126 17:12:43.250611 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-build-system-configs\") pod \"service-telemetry-operator-2-build\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 26 17:12:43 crc kubenswrapper[4856]: I0126 17:12:43.254670 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-builder-dockercfg-8h4xs-push\") pod \"service-telemetry-operator-2-build\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 26 17:12:43 crc kubenswrapper[4856]: I0126 17:12:43.254863 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-builder-dockercfg-8h4xs-pull\") pod \"service-telemetry-operator-2-build\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 26 17:12:43 crc kubenswrapper[4856]: I0126 17:12:43.257641 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spxgc\" (UniqueName: \"kubernetes.io/projected/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-kube-api-access-spxgc\") pod \"service-telemetry-operator-2-build\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " pod="service-telemetry/service-telemetry-operator-2-build" Jan 26 17:12:43 crc kubenswrapper[4856]: I0126 17:12:43.395545 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Jan 26 17:12:46 crc kubenswrapper[4856]: I0126 17:12:46.258660 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Jan 26 17:12:46 crc kubenswrapper[4856]: I0126 17:12:46.729009 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"e3f6dcf4-c152-4a81-8e1d-1fdf469be581","Type":"ContainerStarted","Data":"f38fe849edfe940888da0c7e9589bf8433e33392b1573a13a0d673b63831ce2b"} Jan 26 17:12:49 crc kubenswrapper[4856]: I0126 17:12:49.383783 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-1-build_c90021ae-76d4-45c7-a3b8-04e3b9bbddce/docker-build/0.log" Jan 26 17:12:49 crc kubenswrapper[4856]: I0126 17:12:49.384809 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Jan 26 17:12:49 crc kubenswrapper[4856]: I0126 17:12:49.429240 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-1-build_c90021ae-76d4-45c7-a3b8-04e3b9bbddce/docker-build/0.log" Jan 26 17:12:49 crc kubenswrapper[4856]: I0126 17:12:49.433721 4856 generic.go:334] "Generic (PLEG): container finished" podID="c90021ae-76d4-45c7-a3b8-04e3b9bbddce" containerID="2800b2e70dad7de0c38ad3d8f25f147fcc58b2b450a62480a1e7bc1db9736970" exitCode=1 Jan 26 17:12:49 crc kubenswrapper[4856]: I0126 17:12:49.433767 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"c90021ae-76d4-45c7-a3b8-04e3b9bbddce","Type":"ContainerDied","Data":"2800b2e70dad7de0c38ad3d8f25f147fcc58b2b450a62480a1e7bc1db9736970"} Jan 26 17:12:49 crc kubenswrapper[4856]: I0126 17:12:49.433811 4856 scope.go:117] "RemoveContainer" containerID="2800b2e70dad7de0c38ad3d8f25f147fcc58b2b450a62480a1e7bc1db9736970" Jan 26 17:12:49 crc kubenswrapper[4856]: I0126 17:12:49.453363 4856 scope.go:117] "RemoveContainer" containerID="28aff785d4bd8d892e6c32fa63d73c371a6f2a38a3d0f6ac418a977bacaecdd0" Jan 26 17:12:49 crc kubenswrapper[4856]: I0126 17:12:49.494998 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-builder-dockercfg-8h4xs-pull\") pod \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " Jan 26 17:12:49 crc kubenswrapper[4856]: I0126 17:12:49.495111 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-container-storage-run\") pod \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " Jan 26 17:12:49 crc kubenswrapper[4856]: I0126 17:12:49.495131 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-builder-dockercfg-8h4xs-push\") pod \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " Jan 26 17:12:49 crc kubenswrapper[4856]: I0126 17:12:49.495154 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-container-storage-root\") pod \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " Jan 26 17:12:49 crc kubenswrapper[4856]: I0126 17:12:49.495198 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-build-system-configs\") pod \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " Jan 26 17:12:49 crc kubenswrapper[4856]: I0126 17:12:49.495217 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-buildworkdir\") pod \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " Jan 26 17:12:49 crc kubenswrapper[4856]: I0126 17:12:49.495252 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-buildcachedir\") pod \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " Jan 26 17:12:49 crc kubenswrapper[4856]: I0126 17:12:49.495268 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-build-ca-bundles\") pod \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " Jan 26 17:12:49 crc kubenswrapper[4856]: I0126 17:12:49.495315 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gsm2b\" (UniqueName: \"kubernetes.io/projected/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-kube-api-access-gsm2b\") pod \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " Jan 26 17:12:49 crc kubenswrapper[4856]: I0126 17:12:49.495350 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-build-blob-cache\") pod \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " Jan 26 17:12:49 crc kubenswrapper[4856]: I0126 17:12:49.495364 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-node-pullsecrets\") pod \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " Jan 26 17:12:49 crc kubenswrapper[4856]: I0126 17:12:49.495418 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-build-proxy-ca-bundles\") pod \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\" (UID: \"c90021ae-76d4-45c7-a3b8-04e3b9bbddce\") " Jan 26 17:12:49 crc kubenswrapper[4856]: I0126 17:12:49.495591 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "c90021ae-76d4-45c7-a3b8-04e3b9bbddce" (UID: "c90021ae-76d4-45c7-a3b8-04e3b9bbddce"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:12:49 crc kubenswrapper[4856]: I0126 17:12:49.496008 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "c90021ae-76d4-45c7-a3b8-04e3b9bbddce" (UID: "c90021ae-76d4-45c7-a3b8-04e3b9bbddce"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:12:49 crc kubenswrapper[4856]: I0126 17:12:49.496039 4856 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-buildcachedir\") on node \"crc\" DevicePath \"\"" Jan 26 17:12:49 crc kubenswrapper[4856]: I0126 17:12:49.496154 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "c90021ae-76d4-45c7-a3b8-04e3b9bbddce" (UID: "c90021ae-76d4-45c7-a3b8-04e3b9bbddce"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:12:49 crc kubenswrapper[4856]: I0126 17:12:49.496180 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "c90021ae-76d4-45c7-a3b8-04e3b9bbddce" (UID: "c90021ae-76d4-45c7-a3b8-04e3b9bbddce"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:12:49 crc kubenswrapper[4856]: I0126 17:12:49.496457 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "c90021ae-76d4-45c7-a3b8-04e3b9bbddce" (UID: "c90021ae-76d4-45c7-a3b8-04e3b9bbddce"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:12:49 crc kubenswrapper[4856]: I0126 17:12:49.496874 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "c90021ae-76d4-45c7-a3b8-04e3b9bbddce" (UID: "c90021ae-76d4-45c7-a3b8-04e3b9bbddce"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:12:49 crc kubenswrapper[4856]: I0126 17:12:49.497334 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "c90021ae-76d4-45c7-a3b8-04e3b9bbddce" (UID: "c90021ae-76d4-45c7-a3b8-04e3b9bbddce"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:12:49 crc kubenswrapper[4856]: I0126 17:12:49.497394 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "c90021ae-76d4-45c7-a3b8-04e3b9bbddce" (UID: "c90021ae-76d4-45c7-a3b8-04e3b9bbddce"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:12:49 crc kubenswrapper[4856]: I0126 17:12:49.497733 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "c90021ae-76d4-45c7-a3b8-04e3b9bbddce" (UID: "c90021ae-76d4-45c7-a3b8-04e3b9bbddce"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:12:49 crc kubenswrapper[4856]: I0126 17:12:49.500032 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-builder-dockercfg-8h4xs-pull" (OuterVolumeSpecName: "builder-dockercfg-8h4xs-pull") pod "c90021ae-76d4-45c7-a3b8-04e3b9bbddce" (UID: "c90021ae-76d4-45c7-a3b8-04e3b9bbddce"). InnerVolumeSpecName "builder-dockercfg-8h4xs-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:12:49 crc kubenswrapper[4856]: I0126 17:12:49.500055 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-builder-dockercfg-8h4xs-push" (OuterVolumeSpecName: "builder-dockercfg-8h4xs-push") pod "c90021ae-76d4-45c7-a3b8-04e3b9bbddce" (UID: "c90021ae-76d4-45c7-a3b8-04e3b9bbddce"). InnerVolumeSpecName "builder-dockercfg-8h4xs-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:12:49 crc kubenswrapper[4856]: I0126 17:12:49.500436 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-kube-api-access-gsm2b" (OuterVolumeSpecName: "kube-api-access-gsm2b") pod "c90021ae-76d4-45c7-a3b8-04e3b9bbddce" (UID: "c90021ae-76d4-45c7-a3b8-04e3b9bbddce"). InnerVolumeSpecName "kube-api-access-gsm2b". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:12:49 crc kubenswrapper[4856]: I0126 17:12:49.597474 4856 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-container-storage-run\") on node \"crc\" DevicePath \"\"" Jan 26 17:12:49 crc kubenswrapper[4856]: I0126 17:12:49.597515 4856 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-builder-dockercfg-8h4xs-push\") on node \"crc\" DevicePath \"\"" Jan 26 17:12:49 crc kubenswrapper[4856]: I0126 17:12:49.597574 4856 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-container-storage-root\") on node \"crc\" DevicePath \"\"" Jan 26 17:12:49 crc kubenswrapper[4856]: I0126 17:12:49.597587 4856 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-build-system-configs\") on node \"crc\" DevicePath \"\"" Jan 26 17:12:49 crc kubenswrapper[4856]: I0126 17:12:49.597596 4856 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-buildworkdir\") on node \"crc\" DevicePath \"\"" Jan 26 17:12:49 crc kubenswrapper[4856]: I0126 17:12:49.597604 4856 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 26 17:12:49 crc kubenswrapper[4856]: I0126 17:12:49.597616 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gsm2b\" (UniqueName: \"kubernetes.io/projected/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-kube-api-access-gsm2b\") on node \"crc\" DevicePath \"\"" Jan 26 17:12:49 crc kubenswrapper[4856]: I0126 17:12:49.597624 4856 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-build-blob-cache\") on node \"crc\" DevicePath \"\"" Jan 26 17:12:49 crc kubenswrapper[4856]: I0126 17:12:49.597631 4856 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Jan 26 17:12:49 crc kubenswrapper[4856]: I0126 17:12:49.597638 4856 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 26 17:12:49 crc kubenswrapper[4856]: I0126 17:12:49.597646 4856 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/c90021ae-76d4-45c7-a3b8-04e3b9bbddce-builder-dockercfg-8h4xs-pull\") on node \"crc\" DevicePath \"\"" Jan 26 17:12:50 crc kubenswrapper[4856]: I0126 17:12:50.442154 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"c90021ae-76d4-45c7-a3b8-04e3b9bbddce","Type":"ContainerDied","Data":"dd686a65c0eba5b54c5e815a6d4137b24d38f89f6901d1d1e3907daf3f0a3f46"} Jan 26 17:12:50 crc kubenswrapper[4856]: I0126 17:12:50.442178 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Jan 26 17:12:50 crc kubenswrapper[4856]: I0126 17:12:50.447096 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"e3f6dcf4-c152-4a81-8e1d-1fdf469be581","Type":"ContainerStarted","Data":"afbb877df6ecee92e6604885fb086f3aa9571ea00f867e7fec22ca0c0785ba91"} Jan 26 17:12:50 crc kubenswrapper[4856]: I0126 17:12:50.448080 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"8cba7b0b-8fbc-4d94-a808-43c46e0defaa","Type":"ContainerStarted","Data":"58cea0944e41ac391efcc221b5257ed0cbfcc9d9d995c19c749b68acf88653a0"} Jan 26 17:12:50 crc kubenswrapper[4856]: I0126 17:12:50.535801 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Jan 26 17:12:50 crc kubenswrapper[4856]: I0126 17:12:50.547482 4856 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Jan 26 17:12:51 crc kubenswrapper[4856]: I0126 17:12:51.404856 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c90021ae-76d4-45c7-a3b8-04e3b9bbddce" path="/var/lib/kubelet/pods/c90021ae-76d4-45c7-a3b8-04e3b9bbddce/volumes" Jan 26 17:12:51 crc kubenswrapper[4856]: I0126 17:12:51.458217 4856 generic.go:334] "Generic (PLEG): container finished" podID="8cba7b0b-8fbc-4d94-a808-43c46e0defaa" containerID="58cea0944e41ac391efcc221b5257ed0cbfcc9d9d995c19c749b68acf88653a0" exitCode=0 Jan 26 17:12:51 crc kubenswrapper[4856]: I0126 17:12:51.458626 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"8cba7b0b-8fbc-4d94-a808-43c46e0defaa","Type":"ContainerDied","Data":"58cea0944e41ac391efcc221b5257ed0cbfcc9d9d995c19c749b68acf88653a0"} Jan 26 17:12:52 crc kubenswrapper[4856]: I0126 17:12:52.464885 4856 generic.go:334] "Generic (PLEG): container finished" podID="8cba7b0b-8fbc-4d94-a808-43c46e0defaa" containerID="afc0c6940e36d6454494111916ed37a8c0609122aca51e5a53f48c9728aefbc7" exitCode=0 Jan 26 17:12:52 crc kubenswrapper[4856]: I0126 17:12:52.464946 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"8cba7b0b-8fbc-4d94-a808-43c46e0defaa","Type":"ContainerDied","Data":"afc0c6940e36d6454494111916ed37a8c0609122aca51e5a53f48c9728aefbc7"} Jan 26 17:12:53 crc kubenswrapper[4856]: I0126 17:12:53.474829 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"8cba7b0b-8fbc-4d94-a808-43c46e0defaa","Type":"ContainerStarted","Data":"4a31e616d23d9c76745e34247cf9b597f7d5d4019e0803d40e84b05d4263e2a0"} Jan 26 17:12:53 crc kubenswrapper[4856]: I0126 17:12:53.475556 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:12:53 crc kubenswrapper[4856]: I0126 17:12:53.509977 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/elasticsearch-es-default-0" podStartSLOduration=5.905538427 podStartE2EDuration="58.509936588s" podCreationTimestamp="2026-01-26 17:11:55 +0000 UTC" firstStartedPulling="2026-01-26 17:11:56.569410626 +0000 UTC m=+812.522664607" lastFinishedPulling="2026-01-26 17:12:49.173808787 +0000 UTC m=+865.127062768" observedRunningTime="2026-01-26 17:12:53.503881982 +0000 UTC m=+869.457135963" watchObservedRunningTime="2026-01-26 17:12:53.509936588 +0000 UTC m=+869.463190569" Jan 26 17:12:56 crc kubenswrapper[4856]: I0126 17:12:56.938473 4856 patch_prober.go:28] interesting pod/machine-config-daemon-xm9cq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:12:56 crc kubenswrapper[4856]: I0126 17:12:56.939090 4856 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:12:56 crc kubenswrapper[4856]: I0126 17:12:56.939183 4856 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" Jan 26 17:12:56 crc kubenswrapper[4856]: I0126 17:12:56.940889 4856 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bb3fb578d0ea2b4eb264b402043faa4d1923f5d38749a2ee2c65b084c2e291bd"} pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 26 17:12:56 crc kubenswrapper[4856]: I0126 17:12:56.940980 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" containerName="machine-config-daemon" containerID="cri-o://bb3fb578d0ea2b4eb264b402043faa4d1923f5d38749a2ee2c65b084c2e291bd" gracePeriod=600 Jan 26 17:12:57 crc kubenswrapper[4856]: I0126 17:12:57.559395 4856 generic.go:334] "Generic (PLEG): container finished" podID="63c75ede-5170-4db0-811b-5217ef8d72b3" containerID="bb3fb578d0ea2b4eb264b402043faa4d1923f5d38749a2ee2c65b084c2e291bd" exitCode=0 Jan 26 17:12:57 crc kubenswrapper[4856]: I0126 17:12:57.559473 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" event={"ID":"63c75ede-5170-4db0-811b-5217ef8d72b3","Type":"ContainerDied","Data":"bb3fb578d0ea2b4eb264b402043faa4d1923f5d38749a2ee2c65b084c2e291bd"} Jan 26 17:12:57 crc kubenswrapper[4856]: I0126 17:12:57.559840 4856 scope.go:117] "RemoveContainer" containerID="fe42c0299ac9f35a2260caaf7226f7e2161da013442117dab0d25a7c69c46115" Jan 26 17:12:58 crc kubenswrapper[4856]: I0126 17:12:58.579969 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" event={"ID":"63c75ede-5170-4db0-811b-5217ef8d72b3","Type":"ContainerStarted","Data":"fdaad4602089daad40b0395fbc761e615a8ba2a94c8f5b977142a787034cddb7"} Jan 26 17:12:59 crc kubenswrapper[4856]: I0126 17:12:59.587666 4856 generic.go:334] "Generic (PLEG): container finished" podID="e3f6dcf4-c152-4a81-8e1d-1fdf469be581" containerID="afbb877df6ecee92e6604885fb086f3aa9571ea00f867e7fec22ca0c0785ba91" exitCode=0 Jan 26 17:12:59 crc kubenswrapper[4856]: I0126 17:12:59.587738 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"e3f6dcf4-c152-4a81-8e1d-1fdf469be581","Type":"ContainerDied","Data":"afbb877df6ecee92e6604885fb086f3aa9571ea00f867e7fec22ca0c0785ba91"} Jan 26 17:13:00 crc kubenswrapper[4856]: I0126 17:13:00.595694 4856 generic.go:334] "Generic (PLEG): container finished" podID="e3f6dcf4-c152-4a81-8e1d-1fdf469be581" containerID="8e37eff03e3f55d608de033572142249146d18f2f61d278dcf3a02aa6f6ff2ab" exitCode=0 Jan 26 17:13:00 crc kubenswrapper[4856]: I0126 17:13:00.595762 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"e3f6dcf4-c152-4a81-8e1d-1fdf469be581","Type":"ContainerDied","Data":"8e37eff03e3f55d608de033572142249146d18f2f61d278dcf3a02aa6f6ff2ab"} Jan 26 17:13:00 crc kubenswrapper[4856]: I0126 17:13:00.642007 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-2-build_e3f6dcf4-c152-4a81-8e1d-1fdf469be581/manage-dockerfile/0.log" Jan 26 17:13:01 crc kubenswrapper[4856]: I0126 17:13:01.673673 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"e3f6dcf4-c152-4a81-8e1d-1fdf469be581","Type":"ContainerStarted","Data":"acf8873d5a9fd2dc945aa7f942f92399d79aa34d23d46be85cf69d51f18751c1"} Jan 26 17:13:01 crc kubenswrapper[4856]: I0126 17:13:01.713558 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/service-telemetry-operator-2-build" podStartSLOduration=18.713515358 podStartE2EDuration="18.713515358s" podCreationTimestamp="2026-01-26 17:12:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:13:01.705860629 +0000 UTC m=+877.659114620" watchObservedRunningTime="2026-01-26 17:13:01.713515358 +0000 UTC m=+877.666769339" Jan 26 17:13:06 crc kubenswrapper[4856]: I0126 17:13:06.423186 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="service-telemetry/elasticsearch-es-default-0" podUID="8cba7b0b-8fbc-4d94-a808-43c46e0defaa" containerName="elasticsearch" probeResult="failure" output=< Jan 26 17:13:06 crc kubenswrapper[4856]: {"timestamp": "2026-01-26T17:13:06+00:00", "message": "readiness probe failed", "curl_rc": "7"} Jan 26 17:13:06 crc kubenswrapper[4856]: > Jan 26 17:13:11 crc kubenswrapper[4856]: I0126 17:13:11.252487 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="service-telemetry/elasticsearch-es-default-0" podUID="8cba7b0b-8fbc-4d94-a808-43c46e0defaa" containerName="elasticsearch" probeResult="failure" output=< Jan 26 17:13:11 crc kubenswrapper[4856]: {"timestamp": "2026-01-26T17:13:11+00:00", "message": "readiness probe failed", "curl_rc": "7"} Jan 26 17:13:11 crc kubenswrapper[4856]: > Jan 26 17:13:16 crc kubenswrapper[4856]: I0126 17:13:16.373422 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="service-telemetry/elasticsearch-es-default-0" podUID="8cba7b0b-8fbc-4d94-a808-43c46e0defaa" containerName="elasticsearch" probeResult="failure" output=< Jan 26 17:13:16 crc kubenswrapper[4856]: {"timestamp": "2026-01-26T17:13:16+00:00", "message": "readiness probe failed", "curl_rc": "7"} Jan 26 17:13:16 crc kubenswrapper[4856]: > Jan 26 17:13:21 crc kubenswrapper[4856]: I0126 17:13:21.233422 4856 prober.go:107] "Probe failed" probeType="Readiness" pod="service-telemetry/elasticsearch-es-default-0" podUID="8cba7b0b-8fbc-4d94-a808-43c46e0defaa" containerName="elasticsearch" probeResult="failure" output=< Jan 26 17:13:21 crc kubenswrapper[4856]: {"timestamp": "2026-01-26T17:13:21+00:00", "message": "readiness probe failed", "curl_rc": "7"} Jan 26 17:13:21 crc kubenswrapper[4856]: > Jan 26 17:13:26 crc kubenswrapper[4856]: I0126 17:13:26.344951 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/elasticsearch-es-default-0" Jan 26 17:13:36 crc kubenswrapper[4856]: I0126 17:13:36.978680 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2sbfj"] Jan 26 17:13:36 crc kubenswrapper[4856]: E0126 17:13:36.979476 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c90021ae-76d4-45c7-a3b8-04e3b9bbddce" containerName="docker-build" Jan 26 17:13:36 crc kubenswrapper[4856]: I0126 17:13:36.979494 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="c90021ae-76d4-45c7-a3b8-04e3b9bbddce" containerName="docker-build" Jan 26 17:13:36 crc kubenswrapper[4856]: E0126 17:13:36.979504 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c90021ae-76d4-45c7-a3b8-04e3b9bbddce" containerName="manage-dockerfile" Jan 26 17:13:36 crc kubenswrapper[4856]: I0126 17:13:36.979510 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="c90021ae-76d4-45c7-a3b8-04e3b9bbddce" containerName="manage-dockerfile" Jan 26 17:13:36 crc kubenswrapper[4856]: I0126 17:13:36.979694 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="c90021ae-76d4-45c7-a3b8-04e3b9bbddce" containerName="docker-build" Jan 26 17:13:36 crc kubenswrapper[4856]: I0126 17:13:36.980805 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2sbfj" Jan 26 17:13:36 crc kubenswrapper[4856]: I0126 17:13:36.997263 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2sbfj"] Jan 26 17:13:37 crc kubenswrapper[4856]: I0126 17:13:37.038954 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40ded4e9-2f52-405d-80fb-b4fef311cbc1-catalog-content\") pod \"redhat-operators-2sbfj\" (UID: \"40ded4e9-2f52-405d-80fb-b4fef311cbc1\") " pod="openshift-marketplace/redhat-operators-2sbfj" Jan 26 17:13:37 crc kubenswrapper[4856]: I0126 17:13:37.039182 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40ded4e9-2f52-405d-80fb-b4fef311cbc1-utilities\") pod \"redhat-operators-2sbfj\" (UID: \"40ded4e9-2f52-405d-80fb-b4fef311cbc1\") " pod="openshift-marketplace/redhat-operators-2sbfj" Jan 26 17:13:37 crc kubenswrapper[4856]: I0126 17:13:37.039272 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gl4bs\" (UniqueName: \"kubernetes.io/projected/40ded4e9-2f52-405d-80fb-b4fef311cbc1-kube-api-access-gl4bs\") pod \"redhat-operators-2sbfj\" (UID: \"40ded4e9-2f52-405d-80fb-b4fef311cbc1\") " pod="openshift-marketplace/redhat-operators-2sbfj" Jan 26 17:13:37 crc kubenswrapper[4856]: I0126 17:13:37.140137 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40ded4e9-2f52-405d-80fb-b4fef311cbc1-utilities\") pod \"redhat-operators-2sbfj\" (UID: \"40ded4e9-2f52-405d-80fb-b4fef311cbc1\") " pod="openshift-marketplace/redhat-operators-2sbfj" Jan 26 17:13:37 crc kubenswrapper[4856]: I0126 17:13:37.140222 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gl4bs\" (UniqueName: \"kubernetes.io/projected/40ded4e9-2f52-405d-80fb-b4fef311cbc1-kube-api-access-gl4bs\") pod \"redhat-operators-2sbfj\" (UID: \"40ded4e9-2f52-405d-80fb-b4fef311cbc1\") " pod="openshift-marketplace/redhat-operators-2sbfj" Jan 26 17:13:37 crc kubenswrapper[4856]: I0126 17:13:37.140270 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40ded4e9-2f52-405d-80fb-b4fef311cbc1-catalog-content\") pod \"redhat-operators-2sbfj\" (UID: \"40ded4e9-2f52-405d-80fb-b4fef311cbc1\") " pod="openshift-marketplace/redhat-operators-2sbfj" Jan 26 17:13:37 crc kubenswrapper[4856]: I0126 17:13:37.140935 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40ded4e9-2f52-405d-80fb-b4fef311cbc1-catalog-content\") pod \"redhat-operators-2sbfj\" (UID: \"40ded4e9-2f52-405d-80fb-b4fef311cbc1\") " pod="openshift-marketplace/redhat-operators-2sbfj" Jan 26 17:13:37 crc kubenswrapper[4856]: I0126 17:13:37.140953 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40ded4e9-2f52-405d-80fb-b4fef311cbc1-utilities\") pod \"redhat-operators-2sbfj\" (UID: \"40ded4e9-2f52-405d-80fb-b4fef311cbc1\") " pod="openshift-marketplace/redhat-operators-2sbfj" Jan 26 17:13:37 crc kubenswrapper[4856]: I0126 17:13:37.267083 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gl4bs\" (UniqueName: \"kubernetes.io/projected/40ded4e9-2f52-405d-80fb-b4fef311cbc1-kube-api-access-gl4bs\") pod \"redhat-operators-2sbfj\" (UID: \"40ded4e9-2f52-405d-80fb-b4fef311cbc1\") " pod="openshift-marketplace/redhat-operators-2sbfj" Jan 26 17:13:37 crc kubenswrapper[4856]: I0126 17:13:37.299835 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2sbfj" Jan 26 17:13:38 crc kubenswrapper[4856]: I0126 17:13:38.347765 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2sbfj"] Jan 26 17:13:38 crc kubenswrapper[4856]: W0126 17:13:38.352255 4856 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod40ded4e9_2f52_405d_80fb_b4fef311cbc1.slice/crio-f7181ecfdf3ef508ac91ce2ecf81641c9c7317da5d322beba5c68919dc3a259b WatchSource:0}: Error finding container f7181ecfdf3ef508ac91ce2ecf81641c9c7317da5d322beba5c68919dc3a259b: Status 404 returned error can't find the container with id f7181ecfdf3ef508ac91ce2ecf81641c9c7317da5d322beba5c68919dc3a259b Jan 26 17:13:39 crc kubenswrapper[4856]: I0126 17:13:39.271477 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2sbfj" event={"ID":"40ded4e9-2f52-405d-80fb-b4fef311cbc1","Type":"ContainerStarted","Data":"e5ebb09819be3c66a7ddcceda562a758f2dcc2ef2bb1e57d8921f3dec667e919"} Jan 26 17:13:39 crc kubenswrapper[4856]: I0126 17:13:39.271878 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2sbfj" event={"ID":"40ded4e9-2f52-405d-80fb-b4fef311cbc1","Type":"ContainerStarted","Data":"f7181ecfdf3ef508ac91ce2ecf81641c9c7317da5d322beba5c68919dc3a259b"} Jan 26 17:13:40 crc kubenswrapper[4856]: I0126 17:13:40.279242 4856 generic.go:334] "Generic (PLEG): container finished" podID="40ded4e9-2f52-405d-80fb-b4fef311cbc1" containerID="e5ebb09819be3c66a7ddcceda562a758f2dcc2ef2bb1e57d8921f3dec667e919" exitCode=0 Jan 26 17:13:40 crc kubenswrapper[4856]: I0126 17:13:40.279293 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2sbfj" event={"ID":"40ded4e9-2f52-405d-80fb-b4fef311cbc1","Type":"ContainerDied","Data":"e5ebb09819be3c66a7ddcceda562a758f2dcc2ef2bb1e57d8921f3dec667e919"} Jan 26 17:13:43 crc kubenswrapper[4856]: I0126 17:13:43.313273 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2sbfj" event={"ID":"40ded4e9-2f52-405d-80fb-b4fef311cbc1","Type":"ContainerStarted","Data":"9707c6c51dc5a30aeb3a0d380d791f52a183e2d158ce4607cc75a52d2079c292"} Jan 26 17:13:46 crc kubenswrapper[4856]: I0126 17:13:46.334130 4856 generic.go:334] "Generic (PLEG): container finished" podID="40ded4e9-2f52-405d-80fb-b4fef311cbc1" containerID="9707c6c51dc5a30aeb3a0d380d791f52a183e2d158ce4607cc75a52d2079c292" exitCode=0 Jan 26 17:13:46 crc kubenswrapper[4856]: I0126 17:13:46.334424 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2sbfj" event={"ID":"40ded4e9-2f52-405d-80fb-b4fef311cbc1","Type":"ContainerDied","Data":"9707c6c51dc5a30aeb3a0d380d791f52a183e2d158ce4607cc75a52d2079c292"} Jan 26 17:13:47 crc kubenswrapper[4856]: I0126 17:13:47.468868 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2sbfj" event={"ID":"40ded4e9-2f52-405d-80fb-b4fef311cbc1","Type":"ContainerStarted","Data":"8f51bdcd1b85974530b2a951fcef71cea685b1e499cbb81c32b339883c0f05ca"} Jan 26 17:13:47 crc kubenswrapper[4856]: I0126 17:13:47.552040 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2sbfj" podStartSLOduration=5.094333378 podStartE2EDuration="11.551992493s" podCreationTimestamp="2026-01-26 17:13:36 +0000 UTC" firstStartedPulling="2026-01-26 17:13:40.282435118 +0000 UTC m=+916.235689099" lastFinishedPulling="2026-01-26 17:13:46.740094233 +0000 UTC m=+922.693348214" observedRunningTime="2026-01-26 17:13:47.550957833 +0000 UTC m=+923.504211834" watchObservedRunningTime="2026-01-26 17:13:47.551992493 +0000 UTC m=+923.505246494" Jan 26 17:13:57 crc kubenswrapper[4856]: I0126 17:13:57.299971 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2sbfj" Jan 26 17:13:57 crc kubenswrapper[4856]: I0126 17:13:57.302006 4856 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2sbfj" Jan 26 17:13:57 crc kubenswrapper[4856]: I0126 17:13:57.347324 4856 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2sbfj" Jan 26 17:13:57 crc kubenswrapper[4856]: I0126 17:13:57.778221 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2sbfj" Jan 26 17:13:57 crc kubenswrapper[4856]: I0126 17:13:57.830048 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2sbfj"] Jan 26 17:13:59 crc kubenswrapper[4856]: I0126 17:13:59.750717 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2sbfj" podUID="40ded4e9-2f52-405d-80fb-b4fef311cbc1" containerName="registry-server" containerID="cri-o://8f51bdcd1b85974530b2a951fcef71cea685b1e499cbb81c32b339883c0f05ca" gracePeriod=2 Jan 26 17:14:00 crc kubenswrapper[4856]: I0126 17:14:00.172442 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2sbfj" Jan 26 17:14:00 crc kubenswrapper[4856]: I0126 17:14:00.347053 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gl4bs\" (UniqueName: \"kubernetes.io/projected/40ded4e9-2f52-405d-80fb-b4fef311cbc1-kube-api-access-gl4bs\") pod \"40ded4e9-2f52-405d-80fb-b4fef311cbc1\" (UID: \"40ded4e9-2f52-405d-80fb-b4fef311cbc1\") " Jan 26 17:14:00 crc kubenswrapper[4856]: I0126 17:14:00.347136 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40ded4e9-2f52-405d-80fb-b4fef311cbc1-utilities\") pod \"40ded4e9-2f52-405d-80fb-b4fef311cbc1\" (UID: \"40ded4e9-2f52-405d-80fb-b4fef311cbc1\") " Jan 26 17:14:00 crc kubenswrapper[4856]: I0126 17:14:00.347268 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40ded4e9-2f52-405d-80fb-b4fef311cbc1-catalog-content\") pod \"40ded4e9-2f52-405d-80fb-b4fef311cbc1\" (UID: \"40ded4e9-2f52-405d-80fb-b4fef311cbc1\") " Jan 26 17:14:00 crc kubenswrapper[4856]: I0126 17:14:00.348180 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40ded4e9-2f52-405d-80fb-b4fef311cbc1-utilities" (OuterVolumeSpecName: "utilities") pod "40ded4e9-2f52-405d-80fb-b4fef311cbc1" (UID: "40ded4e9-2f52-405d-80fb-b4fef311cbc1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:14:00 crc kubenswrapper[4856]: I0126 17:14:00.354117 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40ded4e9-2f52-405d-80fb-b4fef311cbc1-kube-api-access-gl4bs" (OuterVolumeSpecName: "kube-api-access-gl4bs") pod "40ded4e9-2f52-405d-80fb-b4fef311cbc1" (UID: "40ded4e9-2f52-405d-80fb-b4fef311cbc1"). InnerVolumeSpecName "kube-api-access-gl4bs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:14:00 crc kubenswrapper[4856]: I0126 17:14:00.449618 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gl4bs\" (UniqueName: \"kubernetes.io/projected/40ded4e9-2f52-405d-80fb-b4fef311cbc1-kube-api-access-gl4bs\") on node \"crc\" DevicePath \"\"" Jan 26 17:14:00 crc kubenswrapper[4856]: I0126 17:14:00.449667 4856 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40ded4e9-2f52-405d-80fb-b4fef311cbc1-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 17:14:00 crc kubenswrapper[4856]: I0126 17:14:00.474875 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40ded4e9-2f52-405d-80fb-b4fef311cbc1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "40ded4e9-2f52-405d-80fb-b4fef311cbc1" (UID: "40ded4e9-2f52-405d-80fb-b4fef311cbc1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:14:00 crc kubenswrapper[4856]: I0126 17:14:00.550830 4856 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40ded4e9-2f52-405d-80fb-b4fef311cbc1-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 17:14:00 crc kubenswrapper[4856]: I0126 17:14:00.760246 4856 generic.go:334] "Generic (PLEG): container finished" podID="40ded4e9-2f52-405d-80fb-b4fef311cbc1" containerID="8f51bdcd1b85974530b2a951fcef71cea685b1e499cbb81c32b339883c0f05ca" exitCode=0 Jan 26 17:14:00 crc kubenswrapper[4856]: I0126 17:14:00.761266 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2sbfj" event={"ID":"40ded4e9-2f52-405d-80fb-b4fef311cbc1","Type":"ContainerDied","Data":"8f51bdcd1b85974530b2a951fcef71cea685b1e499cbb81c32b339883c0f05ca"} Jan 26 17:14:00 crc kubenswrapper[4856]: I0126 17:14:00.761381 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2sbfj" event={"ID":"40ded4e9-2f52-405d-80fb-b4fef311cbc1","Type":"ContainerDied","Data":"f7181ecfdf3ef508ac91ce2ecf81641c9c7317da5d322beba5c68919dc3a259b"} Jan 26 17:14:00 crc kubenswrapper[4856]: I0126 17:14:00.761500 4856 scope.go:117] "RemoveContainer" containerID="8f51bdcd1b85974530b2a951fcef71cea685b1e499cbb81c32b339883c0f05ca" Jan 26 17:14:00 crc kubenswrapper[4856]: I0126 17:14:00.761726 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2sbfj" Jan 26 17:14:00 crc kubenswrapper[4856]: I0126 17:14:00.784330 4856 scope.go:117] "RemoveContainer" containerID="9707c6c51dc5a30aeb3a0d380d791f52a183e2d158ce4607cc75a52d2079c292" Jan 26 17:14:00 crc kubenswrapper[4856]: I0126 17:14:00.797777 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2sbfj"] Jan 26 17:14:00 crc kubenswrapper[4856]: I0126 17:14:00.804563 4856 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2sbfj"] Jan 26 17:14:00 crc kubenswrapper[4856]: I0126 17:14:00.824026 4856 scope.go:117] "RemoveContainer" containerID="e5ebb09819be3c66a7ddcceda562a758f2dcc2ef2bb1e57d8921f3dec667e919" Jan 26 17:14:00 crc kubenswrapper[4856]: I0126 17:14:00.838709 4856 scope.go:117] "RemoveContainer" containerID="8f51bdcd1b85974530b2a951fcef71cea685b1e499cbb81c32b339883c0f05ca" Jan 26 17:14:00 crc kubenswrapper[4856]: E0126 17:14:00.839184 4856 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f51bdcd1b85974530b2a951fcef71cea685b1e499cbb81c32b339883c0f05ca\": container with ID starting with 8f51bdcd1b85974530b2a951fcef71cea685b1e499cbb81c32b339883c0f05ca not found: ID does not exist" containerID="8f51bdcd1b85974530b2a951fcef71cea685b1e499cbb81c32b339883c0f05ca" Jan 26 17:14:00 crc kubenswrapper[4856]: I0126 17:14:00.839260 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f51bdcd1b85974530b2a951fcef71cea685b1e499cbb81c32b339883c0f05ca"} err="failed to get container status \"8f51bdcd1b85974530b2a951fcef71cea685b1e499cbb81c32b339883c0f05ca\": rpc error: code = NotFound desc = could not find container \"8f51bdcd1b85974530b2a951fcef71cea685b1e499cbb81c32b339883c0f05ca\": container with ID starting with 8f51bdcd1b85974530b2a951fcef71cea685b1e499cbb81c32b339883c0f05ca not found: ID does not exist" Jan 26 17:14:00 crc kubenswrapper[4856]: I0126 17:14:00.839302 4856 scope.go:117] "RemoveContainer" containerID="9707c6c51dc5a30aeb3a0d380d791f52a183e2d158ce4607cc75a52d2079c292" Jan 26 17:14:00 crc kubenswrapper[4856]: E0126 17:14:00.839659 4856 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9707c6c51dc5a30aeb3a0d380d791f52a183e2d158ce4607cc75a52d2079c292\": container with ID starting with 9707c6c51dc5a30aeb3a0d380d791f52a183e2d158ce4607cc75a52d2079c292 not found: ID does not exist" containerID="9707c6c51dc5a30aeb3a0d380d791f52a183e2d158ce4607cc75a52d2079c292" Jan 26 17:14:00 crc kubenswrapper[4856]: I0126 17:14:00.839692 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9707c6c51dc5a30aeb3a0d380d791f52a183e2d158ce4607cc75a52d2079c292"} err="failed to get container status \"9707c6c51dc5a30aeb3a0d380d791f52a183e2d158ce4607cc75a52d2079c292\": rpc error: code = NotFound desc = could not find container \"9707c6c51dc5a30aeb3a0d380d791f52a183e2d158ce4607cc75a52d2079c292\": container with ID starting with 9707c6c51dc5a30aeb3a0d380d791f52a183e2d158ce4607cc75a52d2079c292 not found: ID does not exist" Jan 26 17:14:00 crc kubenswrapper[4856]: I0126 17:14:00.839715 4856 scope.go:117] "RemoveContainer" containerID="e5ebb09819be3c66a7ddcceda562a758f2dcc2ef2bb1e57d8921f3dec667e919" Jan 26 17:14:00 crc kubenswrapper[4856]: E0126 17:14:00.839989 4856 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5ebb09819be3c66a7ddcceda562a758f2dcc2ef2bb1e57d8921f3dec667e919\": container with ID starting with e5ebb09819be3c66a7ddcceda562a758f2dcc2ef2bb1e57d8921f3dec667e919 not found: ID does not exist" containerID="e5ebb09819be3c66a7ddcceda562a758f2dcc2ef2bb1e57d8921f3dec667e919" Jan 26 17:14:00 crc kubenswrapper[4856]: I0126 17:14:00.840019 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5ebb09819be3c66a7ddcceda562a758f2dcc2ef2bb1e57d8921f3dec667e919"} err="failed to get container status \"e5ebb09819be3c66a7ddcceda562a758f2dcc2ef2bb1e57d8921f3dec667e919\": rpc error: code = NotFound desc = could not find container \"e5ebb09819be3c66a7ddcceda562a758f2dcc2ef2bb1e57d8921f3dec667e919\": container with ID starting with e5ebb09819be3c66a7ddcceda562a758f2dcc2ef2bb1e57d8921f3dec667e919 not found: ID does not exist" Jan 26 17:14:01 crc kubenswrapper[4856]: I0126 17:14:01.405043 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40ded4e9-2f52-405d-80fb-b4fef311cbc1" path="/var/lib/kubelet/pods/40ded4e9-2f52-405d-80fb-b4fef311cbc1/volumes" Jan 26 17:14:03 crc kubenswrapper[4856]: I0126 17:14:03.384806 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wfmp4"] Jan 26 17:14:03 crc kubenswrapper[4856]: E0126 17:14:03.385191 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40ded4e9-2f52-405d-80fb-b4fef311cbc1" containerName="extract-utilities" Jan 26 17:14:03 crc kubenswrapper[4856]: I0126 17:14:03.385213 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="40ded4e9-2f52-405d-80fb-b4fef311cbc1" containerName="extract-utilities" Jan 26 17:14:03 crc kubenswrapper[4856]: E0126 17:14:03.385227 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40ded4e9-2f52-405d-80fb-b4fef311cbc1" containerName="registry-server" Jan 26 17:14:03 crc kubenswrapper[4856]: I0126 17:14:03.385236 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="40ded4e9-2f52-405d-80fb-b4fef311cbc1" containerName="registry-server" Jan 26 17:14:03 crc kubenswrapper[4856]: E0126 17:14:03.385250 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40ded4e9-2f52-405d-80fb-b4fef311cbc1" containerName="extract-content" Jan 26 17:14:03 crc kubenswrapper[4856]: I0126 17:14:03.385256 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="40ded4e9-2f52-405d-80fb-b4fef311cbc1" containerName="extract-content" Jan 26 17:14:03 crc kubenswrapper[4856]: I0126 17:14:03.385400 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="40ded4e9-2f52-405d-80fb-b4fef311cbc1" containerName="registry-server" Jan 26 17:14:03 crc kubenswrapper[4856]: I0126 17:14:03.386401 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wfmp4" Jan 26 17:14:03 crc kubenswrapper[4856]: I0126 17:14:03.404764 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wfmp4"] Jan 26 17:14:03 crc kubenswrapper[4856]: I0126 17:14:03.486806 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce-catalog-content\") pod \"community-operators-wfmp4\" (UID: \"c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce\") " pod="openshift-marketplace/community-operators-wfmp4" Jan 26 17:14:03 crc kubenswrapper[4856]: I0126 17:14:03.486865 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce-utilities\") pod \"community-operators-wfmp4\" (UID: \"c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce\") " pod="openshift-marketplace/community-operators-wfmp4" Jan 26 17:14:03 crc kubenswrapper[4856]: I0126 17:14:03.486902 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rq2j6\" (UniqueName: \"kubernetes.io/projected/c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce-kube-api-access-rq2j6\") pod \"community-operators-wfmp4\" (UID: \"c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce\") " pod="openshift-marketplace/community-operators-wfmp4" Jan 26 17:14:03 crc kubenswrapper[4856]: I0126 17:14:03.588134 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rq2j6\" (UniqueName: \"kubernetes.io/projected/c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce-kube-api-access-rq2j6\") pod \"community-operators-wfmp4\" (UID: \"c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce\") " pod="openshift-marketplace/community-operators-wfmp4" Jan 26 17:14:03 crc kubenswrapper[4856]: I0126 17:14:03.588671 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce-catalog-content\") pod \"community-operators-wfmp4\" (UID: \"c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce\") " pod="openshift-marketplace/community-operators-wfmp4" Jan 26 17:14:03 crc kubenswrapper[4856]: I0126 17:14:03.588724 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce-utilities\") pod \"community-operators-wfmp4\" (UID: \"c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce\") " pod="openshift-marketplace/community-operators-wfmp4" Jan 26 17:14:03 crc kubenswrapper[4856]: I0126 17:14:03.589225 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce-utilities\") pod \"community-operators-wfmp4\" (UID: \"c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce\") " pod="openshift-marketplace/community-operators-wfmp4" Jan 26 17:14:03 crc kubenswrapper[4856]: I0126 17:14:03.589346 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce-catalog-content\") pod \"community-operators-wfmp4\" (UID: \"c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce\") " pod="openshift-marketplace/community-operators-wfmp4" Jan 26 17:14:03 crc kubenswrapper[4856]: I0126 17:14:03.611670 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rq2j6\" (UniqueName: \"kubernetes.io/projected/c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce-kube-api-access-rq2j6\") pod \"community-operators-wfmp4\" (UID: \"c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce\") " pod="openshift-marketplace/community-operators-wfmp4" Jan 26 17:14:03 crc kubenswrapper[4856]: I0126 17:14:03.714335 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wfmp4" Jan 26 17:14:03 crc kubenswrapper[4856]: I0126 17:14:03.960125 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wfmp4"] Jan 26 17:14:04 crc kubenswrapper[4856]: I0126 17:14:04.793132 4856 generic.go:334] "Generic (PLEG): container finished" podID="c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce" containerID="f7fbcfd835618d9f317d705cee22aec914f20459f0e87d10fdb7999dbc362e73" exitCode=0 Jan 26 17:14:04 crc kubenswrapper[4856]: I0126 17:14:04.793200 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wfmp4" event={"ID":"c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce","Type":"ContainerDied","Data":"f7fbcfd835618d9f317d705cee22aec914f20459f0e87d10fdb7999dbc362e73"} Jan 26 17:14:04 crc kubenswrapper[4856]: I0126 17:14:04.793227 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wfmp4" event={"ID":"c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce","Type":"ContainerStarted","Data":"5fe590faf2a2e22175bca3437f2f8e5267371a1d72628c7cdbbefa6d6d6e3e3a"} Jan 26 17:14:05 crc kubenswrapper[4856]: I0126 17:14:05.801683 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wfmp4" event={"ID":"c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce","Type":"ContainerStarted","Data":"4fbeea5f636f158b21305e61d7dfa24c3c339ebddc6eb377deb62b3a93364006"} Jan 26 17:14:06 crc kubenswrapper[4856]: I0126 17:14:06.810939 4856 generic.go:334] "Generic (PLEG): container finished" podID="c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce" containerID="4fbeea5f636f158b21305e61d7dfa24c3c339ebddc6eb377deb62b3a93364006" exitCode=0 Jan 26 17:14:06 crc kubenswrapper[4856]: I0126 17:14:06.811069 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wfmp4" event={"ID":"c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce","Type":"ContainerDied","Data":"4fbeea5f636f158b21305e61d7dfa24c3c339ebddc6eb377deb62b3a93364006"} Jan 26 17:14:08 crc kubenswrapper[4856]: I0126 17:14:08.847410 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wfmp4" event={"ID":"c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce","Type":"ContainerStarted","Data":"77a7df0f62462fb40d0e58d76dda48ebc4b0ee8b1269870dc7102ccd1a2e6a09"} Jan 26 17:14:08 crc kubenswrapper[4856]: I0126 17:14:08.870955 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wfmp4" podStartSLOduration=3.191299689 podStartE2EDuration="5.870904191s" podCreationTimestamp="2026-01-26 17:14:03 +0000 UTC" firstStartedPulling="2026-01-26 17:14:04.795085118 +0000 UTC m=+940.748339099" lastFinishedPulling="2026-01-26 17:14:07.4746896 +0000 UTC m=+943.427943601" observedRunningTime="2026-01-26 17:14:08.870741596 +0000 UTC m=+944.823995587" watchObservedRunningTime="2026-01-26 17:14:08.870904191 +0000 UTC m=+944.824158172" Jan 26 17:14:13 crc kubenswrapper[4856]: I0126 17:14:13.715107 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wfmp4" Jan 26 17:14:13 crc kubenswrapper[4856]: I0126 17:14:13.715394 4856 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wfmp4" Jan 26 17:14:13 crc kubenswrapper[4856]: I0126 17:14:13.756821 4856 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wfmp4" Jan 26 17:14:13 crc kubenswrapper[4856]: I0126 17:14:13.994746 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wfmp4" Jan 26 17:14:14 crc kubenswrapper[4856]: I0126 17:14:14.047379 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wfmp4"] Jan 26 17:14:15 crc kubenswrapper[4856]: I0126 17:14:15.909108 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wfmp4" podUID="c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce" containerName="registry-server" containerID="cri-o://77a7df0f62462fb40d0e58d76dda48ebc4b0ee8b1269870dc7102ccd1a2e6a09" gracePeriod=2 Jan 26 17:14:19 crc kubenswrapper[4856]: I0126 17:14:19.870424 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wfmp4" Jan 26 17:14:19 crc kubenswrapper[4856]: I0126 17:14:19.941887 4856 generic.go:334] "Generic (PLEG): container finished" podID="c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce" containerID="77a7df0f62462fb40d0e58d76dda48ebc4b0ee8b1269870dc7102ccd1a2e6a09" exitCode=0 Jan 26 17:14:19 crc kubenswrapper[4856]: I0126 17:14:19.941934 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wfmp4" event={"ID":"c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce","Type":"ContainerDied","Data":"77a7df0f62462fb40d0e58d76dda48ebc4b0ee8b1269870dc7102ccd1a2e6a09"} Jan 26 17:14:19 crc kubenswrapper[4856]: I0126 17:14:19.941950 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wfmp4" Jan 26 17:14:19 crc kubenswrapper[4856]: I0126 17:14:19.941976 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wfmp4" event={"ID":"c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce","Type":"ContainerDied","Data":"5fe590faf2a2e22175bca3437f2f8e5267371a1d72628c7cdbbefa6d6d6e3e3a"} Jan 26 17:14:19 crc kubenswrapper[4856]: I0126 17:14:19.942030 4856 scope.go:117] "RemoveContainer" containerID="77a7df0f62462fb40d0e58d76dda48ebc4b0ee8b1269870dc7102ccd1a2e6a09" Jan 26 17:14:19 crc kubenswrapper[4856]: I0126 17:14:19.954051 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rq2j6\" (UniqueName: \"kubernetes.io/projected/c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce-kube-api-access-rq2j6\") pod \"c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce\" (UID: \"c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce\") " Jan 26 17:14:19 crc kubenswrapper[4856]: I0126 17:14:19.954299 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce-catalog-content\") pod \"c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce\" (UID: \"c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce\") " Jan 26 17:14:19 crc kubenswrapper[4856]: I0126 17:14:19.954333 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce-utilities\") pod \"c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce\" (UID: \"c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce\") " Jan 26 17:14:19 crc kubenswrapper[4856]: I0126 17:14:19.955256 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce-utilities" (OuterVolumeSpecName: "utilities") pod "c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce" (UID: "c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:14:19 crc kubenswrapper[4856]: I0126 17:14:19.960741 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce-kube-api-access-rq2j6" (OuterVolumeSpecName: "kube-api-access-rq2j6") pod "c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce" (UID: "c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce"). InnerVolumeSpecName "kube-api-access-rq2j6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:14:19 crc kubenswrapper[4856]: I0126 17:14:19.961853 4856 scope.go:117] "RemoveContainer" containerID="4fbeea5f636f158b21305e61d7dfa24c3c339ebddc6eb377deb62b3a93364006" Jan 26 17:14:20 crc kubenswrapper[4856]: I0126 17:14:20.002335 4856 scope.go:117] "RemoveContainer" containerID="f7fbcfd835618d9f317d705cee22aec914f20459f0e87d10fdb7999dbc362e73" Jan 26 17:14:20 crc kubenswrapper[4856]: I0126 17:14:20.015237 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce" (UID: "c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:14:20 crc kubenswrapper[4856]: I0126 17:14:20.024788 4856 scope.go:117] "RemoveContainer" containerID="77a7df0f62462fb40d0e58d76dda48ebc4b0ee8b1269870dc7102ccd1a2e6a09" Jan 26 17:14:20 crc kubenswrapper[4856]: E0126 17:14:20.025961 4856 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77a7df0f62462fb40d0e58d76dda48ebc4b0ee8b1269870dc7102ccd1a2e6a09\": container with ID starting with 77a7df0f62462fb40d0e58d76dda48ebc4b0ee8b1269870dc7102ccd1a2e6a09 not found: ID does not exist" containerID="77a7df0f62462fb40d0e58d76dda48ebc4b0ee8b1269870dc7102ccd1a2e6a09" Jan 26 17:14:20 crc kubenswrapper[4856]: I0126 17:14:20.026016 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77a7df0f62462fb40d0e58d76dda48ebc4b0ee8b1269870dc7102ccd1a2e6a09"} err="failed to get container status \"77a7df0f62462fb40d0e58d76dda48ebc4b0ee8b1269870dc7102ccd1a2e6a09\": rpc error: code = NotFound desc = could not find container \"77a7df0f62462fb40d0e58d76dda48ebc4b0ee8b1269870dc7102ccd1a2e6a09\": container with ID starting with 77a7df0f62462fb40d0e58d76dda48ebc4b0ee8b1269870dc7102ccd1a2e6a09 not found: ID does not exist" Jan 26 17:14:20 crc kubenswrapper[4856]: I0126 17:14:20.026045 4856 scope.go:117] "RemoveContainer" containerID="4fbeea5f636f158b21305e61d7dfa24c3c339ebddc6eb377deb62b3a93364006" Jan 26 17:14:20 crc kubenswrapper[4856]: E0126 17:14:20.026273 4856 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4fbeea5f636f158b21305e61d7dfa24c3c339ebddc6eb377deb62b3a93364006\": container with ID starting with 4fbeea5f636f158b21305e61d7dfa24c3c339ebddc6eb377deb62b3a93364006 not found: ID does not exist" containerID="4fbeea5f636f158b21305e61d7dfa24c3c339ebddc6eb377deb62b3a93364006" Jan 26 17:14:20 crc kubenswrapper[4856]: I0126 17:14:20.026299 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fbeea5f636f158b21305e61d7dfa24c3c339ebddc6eb377deb62b3a93364006"} err="failed to get container status \"4fbeea5f636f158b21305e61d7dfa24c3c339ebddc6eb377deb62b3a93364006\": rpc error: code = NotFound desc = could not find container \"4fbeea5f636f158b21305e61d7dfa24c3c339ebddc6eb377deb62b3a93364006\": container with ID starting with 4fbeea5f636f158b21305e61d7dfa24c3c339ebddc6eb377deb62b3a93364006 not found: ID does not exist" Jan 26 17:14:20 crc kubenswrapper[4856]: I0126 17:14:20.026312 4856 scope.go:117] "RemoveContainer" containerID="f7fbcfd835618d9f317d705cee22aec914f20459f0e87d10fdb7999dbc362e73" Jan 26 17:14:20 crc kubenswrapper[4856]: E0126 17:14:20.028018 4856 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7fbcfd835618d9f317d705cee22aec914f20459f0e87d10fdb7999dbc362e73\": container with ID starting with f7fbcfd835618d9f317d705cee22aec914f20459f0e87d10fdb7999dbc362e73 not found: ID does not exist" containerID="f7fbcfd835618d9f317d705cee22aec914f20459f0e87d10fdb7999dbc362e73" Jan 26 17:14:20 crc kubenswrapper[4856]: I0126 17:14:20.028048 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7fbcfd835618d9f317d705cee22aec914f20459f0e87d10fdb7999dbc362e73"} err="failed to get container status \"f7fbcfd835618d9f317d705cee22aec914f20459f0e87d10fdb7999dbc362e73\": rpc error: code = NotFound desc = could not find container \"f7fbcfd835618d9f317d705cee22aec914f20459f0e87d10fdb7999dbc362e73\": container with ID starting with f7fbcfd835618d9f317d705cee22aec914f20459f0e87d10fdb7999dbc362e73 not found: ID does not exist" Jan 26 17:14:20 crc kubenswrapper[4856]: I0126 17:14:20.056635 4856 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 17:14:20 crc kubenswrapper[4856]: I0126 17:14:20.056685 4856 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 17:14:20 crc kubenswrapper[4856]: I0126 17:14:20.056695 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rq2j6\" (UniqueName: \"kubernetes.io/projected/c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce-kube-api-access-rq2j6\") on node \"crc\" DevicePath \"\"" Jan 26 17:14:20 crc kubenswrapper[4856]: I0126 17:14:20.275004 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wfmp4"] Jan 26 17:14:20 crc kubenswrapper[4856]: I0126 17:14:20.288226 4856 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wfmp4"] Jan 26 17:14:21 crc kubenswrapper[4856]: I0126 17:14:21.403217 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce" path="/var/lib/kubelet/pods/c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce/volumes" Jan 26 17:15:00 crc kubenswrapper[4856]: I0126 17:15:00.165510 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490795-qbctw"] Jan 26 17:15:00 crc kubenswrapper[4856]: E0126 17:15:00.166723 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce" containerName="extract-content" Jan 26 17:15:00 crc kubenswrapper[4856]: I0126 17:15:00.166761 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce" containerName="extract-content" Jan 26 17:15:00 crc kubenswrapper[4856]: E0126 17:15:00.166778 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce" containerName="extract-utilities" Jan 26 17:15:00 crc kubenswrapper[4856]: I0126 17:15:00.166790 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce" containerName="extract-utilities" Jan 26 17:15:00 crc kubenswrapper[4856]: E0126 17:15:00.166837 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce" containerName="registry-server" Jan 26 17:15:00 crc kubenswrapper[4856]: I0126 17:15:00.166857 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce" containerName="registry-server" Jan 26 17:15:00 crc kubenswrapper[4856]: I0126 17:15:00.167120 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0fe9943-bcf5-4b7d-b093-5ee2e453b2ce" containerName="registry-server" Jan 26 17:15:00 crc kubenswrapper[4856]: I0126 17:15:00.168051 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490795-qbctw" Jan 26 17:15:00 crc kubenswrapper[4856]: I0126 17:15:00.170114 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 26 17:15:00 crc kubenswrapper[4856]: I0126 17:15:00.171669 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 26 17:15:00 crc kubenswrapper[4856]: I0126 17:15:00.177281 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490795-qbctw"] Jan 26 17:15:00 crc kubenswrapper[4856]: I0126 17:15:00.352655 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b3f61a40-5427-4fe8-89d2-92b71f9e1052-config-volume\") pod \"collect-profiles-29490795-qbctw\" (UID: \"b3f61a40-5427-4fe8-89d2-92b71f9e1052\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490795-qbctw" Jan 26 17:15:00 crc kubenswrapper[4856]: I0126 17:15:00.352740 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b3f61a40-5427-4fe8-89d2-92b71f9e1052-secret-volume\") pod \"collect-profiles-29490795-qbctw\" (UID: \"b3f61a40-5427-4fe8-89d2-92b71f9e1052\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490795-qbctw" Jan 26 17:15:00 crc kubenswrapper[4856]: I0126 17:15:00.352811 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxp5k\" (UniqueName: \"kubernetes.io/projected/b3f61a40-5427-4fe8-89d2-92b71f9e1052-kube-api-access-vxp5k\") pod \"collect-profiles-29490795-qbctw\" (UID: \"b3f61a40-5427-4fe8-89d2-92b71f9e1052\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490795-qbctw" Jan 26 17:15:00 crc kubenswrapper[4856]: I0126 17:15:00.455337 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b3f61a40-5427-4fe8-89d2-92b71f9e1052-config-volume\") pod \"collect-profiles-29490795-qbctw\" (UID: \"b3f61a40-5427-4fe8-89d2-92b71f9e1052\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490795-qbctw" Jan 26 17:15:00 crc kubenswrapper[4856]: I0126 17:15:00.455386 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b3f61a40-5427-4fe8-89d2-92b71f9e1052-secret-volume\") pod \"collect-profiles-29490795-qbctw\" (UID: \"b3f61a40-5427-4fe8-89d2-92b71f9e1052\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490795-qbctw" Jan 26 17:15:00 crc kubenswrapper[4856]: I0126 17:15:00.455422 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxp5k\" (UniqueName: \"kubernetes.io/projected/b3f61a40-5427-4fe8-89d2-92b71f9e1052-kube-api-access-vxp5k\") pod \"collect-profiles-29490795-qbctw\" (UID: \"b3f61a40-5427-4fe8-89d2-92b71f9e1052\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490795-qbctw" Jan 26 17:15:00 crc kubenswrapper[4856]: I0126 17:15:00.456390 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b3f61a40-5427-4fe8-89d2-92b71f9e1052-config-volume\") pod \"collect-profiles-29490795-qbctw\" (UID: \"b3f61a40-5427-4fe8-89d2-92b71f9e1052\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490795-qbctw" Jan 26 17:15:00 crc kubenswrapper[4856]: I0126 17:15:00.461650 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b3f61a40-5427-4fe8-89d2-92b71f9e1052-secret-volume\") pod \"collect-profiles-29490795-qbctw\" (UID: \"b3f61a40-5427-4fe8-89d2-92b71f9e1052\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490795-qbctw" Jan 26 17:15:00 crc kubenswrapper[4856]: I0126 17:15:00.471998 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxp5k\" (UniqueName: \"kubernetes.io/projected/b3f61a40-5427-4fe8-89d2-92b71f9e1052-kube-api-access-vxp5k\") pod \"collect-profiles-29490795-qbctw\" (UID: \"b3f61a40-5427-4fe8-89d2-92b71f9e1052\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490795-qbctw" Jan 26 17:15:00 crc kubenswrapper[4856]: I0126 17:15:00.491849 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490795-qbctw" Jan 26 17:15:00 crc kubenswrapper[4856]: I0126 17:15:00.905263 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490795-qbctw"] Jan 26 17:15:01 crc kubenswrapper[4856]: I0126 17:15:01.265458 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490795-qbctw" event={"ID":"b3f61a40-5427-4fe8-89d2-92b71f9e1052","Type":"ContainerStarted","Data":"3f0f867caa951dd961753ccea7d1191209079f2d8ed4c2ee8073b09d12c6f2aa"} Jan 26 17:15:02 crc kubenswrapper[4856]: I0126 17:15:02.274721 4856 generic.go:334] "Generic (PLEG): container finished" podID="b3f61a40-5427-4fe8-89d2-92b71f9e1052" containerID="01b95cef7e01f8893038633345035674376d741feceba3066786508454c91270" exitCode=0 Jan 26 17:15:02 crc kubenswrapper[4856]: I0126 17:15:02.275048 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490795-qbctw" event={"ID":"b3f61a40-5427-4fe8-89d2-92b71f9e1052","Type":"ContainerDied","Data":"01b95cef7e01f8893038633345035674376d741feceba3066786508454c91270"} Jan 26 17:15:03 crc kubenswrapper[4856]: I0126 17:15:03.597620 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490795-qbctw" Jan 26 17:15:03 crc kubenswrapper[4856]: I0126 17:15:03.797593 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b3f61a40-5427-4fe8-89d2-92b71f9e1052-config-volume\") pod \"b3f61a40-5427-4fe8-89d2-92b71f9e1052\" (UID: \"b3f61a40-5427-4fe8-89d2-92b71f9e1052\") " Jan 26 17:15:03 crc kubenswrapper[4856]: I0126 17:15:03.797643 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxp5k\" (UniqueName: \"kubernetes.io/projected/b3f61a40-5427-4fe8-89d2-92b71f9e1052-kube-api-access-vxp5k\") pod \"b3f61a40-5427-4fe8-89d2-92b71f9e1052\" (UID: \"b3f61a40-5427-4fe8-89d2-92b71f9e1052\") " Jan 26 17:15:03 crc kubenswrapper[4856]: I0126 17:15:03.797774 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b3f61a40-5427-4fe8-89d2-92b71f9e1052-secret-volume\") pod \"b3f61a40-5427-4fe8-89d2-92b71f9e1052\" (UID: \"b3f61a40-5427-4fe8-89d2-92b71f9e1052\") " Jan 26 17:15:03 crc kubenswrapper[4856]: I0126 17:15:03.798569 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3f61a40-5427-4fe8-89d2-92b71f9e1052-config-volume" (OuterVolumeSpecName: "config-volume") pod "b3f61a40-5427-4fe8-89d2-92b71f9e1052" (UID: "b3f61a40-5427-4fe8-89d2-92b71f9e1052"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:15:03 crc kubenswrapper[4856]: I0126 17:15:03.803482 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3f61a40-5427-4fe8-89d2-92b71f9e1052-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b3f61a40-5427-4fe8-89d2-92b71f9e1052" (UID: "b3f61a40-5427-4fe8-89d2-92b71f9e1052"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:15:03 crc kubenswrapper[4856]: I0126 17:15:03.804163 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3f61a40-5427-4fe8-89d2-92b71f9e1052-kube-api-access-vxp5k" (OuterVolumeSpecName: "kube-api-access-vxp5k") pod "b3f61a40-5427-4fe8-89d2-92b71f9e1052" (UID: "b3f61a40-5427-4fe8-89d2-92b71f9e1052"). InnerVolumeSpecName "kube-api-access-vxp5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:15:03 crc kubenswrapper[4856]: I0126 17:15:03.899026 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxp5k\" (UniqueName: \"kubernetes.io/projected/b3f61a40-5427-4fe8-89d2-92b71f9e1052-kube-api-access-vxp5k\") on node \"crc\" DevicePath \"\"" Jan 26 17:15:03 crc kubenswrapper[4856]: I0126 17:15:03.899063 4856 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b3f61a40-5427-4fe8-89d2-92b71f9e1052-config-volume\") on node \"crc\" DevicePath \"\"" Jan 26 17:15:03 crc kubenswrapper[4856]: I0126 17:15:03.899075 4856 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b3f61a40-5427-4fe8-89d2-92b71f9e1052-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 26 17:15:04 crc kubenswrapper[4856]: I0126 17:15:04.291669 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490795-qbctw" event={"ID":"b3f61a40-5427-4fe8-89d2-92b71f9e1052","Type":"ContainerDied","Data":"3f0f867caa951dd961753ccea7d1191209079f2d8ed4c2ee8073b09d12c6f2aa"} Jan 26 17:15:04 crc kubenswrapper[4856]: I0126 17:15:04.291725 4856 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f0f867caa951dd961753ccea7d1191209079f2d8ed4c2ee8073b09d12c6f2aa" Jan 26 17:15:04 crc kubenswrapper[4856]: I0126 17:15:04.292231 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490795-qbctw" Jan 26 17:15:08 crc kubenswrapper[4856]: I0126 17:15:08.318004 4856 generic.go:334] "Generic (PLEG): container finished" podID="e3f6dcf4-c152-4a81-8e1d-1fdf469be581" containerID="acf8873d5a9fd2dc945aa7f942f92399d79aa34d23d46be85cf69d51f18751c1" exitCode=0 Jan 26 17:15:08 crc kubenswrapper[4856]: I0126 17:15:08.318065 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"e3f6dcf4-c152-4a81-8e1d-1fdf469be581","Type":"ContainerDied","Data":"acf8873d5a9fd2dc945aa7f942f92399d79aa34d23d46be85cf69d51f18751c1"} Jan 26 17:15:09 crc kubenswrapper[4856]: I0126 17:15:09.618782 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Jan 26 17:15:09 crc kubenswrapper[4856]: I0126 17:15:09.678689 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-buildcachedir\") pod \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " Jan 26 17:15:09 crc kubenswrapper[4856]: I0126 17:15:09.678830 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-build-proxy-ca-bundles\") pod \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " Jan 26 17:15:09 crc kubenswrapper[4856]: I0126 17:15:09.678840 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "e3f6dcf4-c152-4a81-8e1d-1fdf469be581" (UID: "e3f6dcf4-c152-4a81-8e1d-1fdf469be581"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:15:09 crc kubenswrapper[4856]: I0126 17:15:09.678865 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-build-system-configs\") pod \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " Jan 26 17:15:09 crc kubenswrapper[4856]: I0126 17:15:09.678940 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-build-blob-cache\") pod \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " Jan 26 17:15:09 crc kubenswrapper[4856]: I0126 17:15:09.679003 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-buildworkdir\") pod \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " Jan 26 17:15:09 crc kubenswrapper[4856]: I0126 17:15:09.679025 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-spxgc\" (UniqueName: \"kubernetes.io/projected/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-kube-api-access-spxgc\") pod \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " Jan 26 17:15:09 crc kubenswrapper[4856]: I0126 17:15:09.679048 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-container-storage-root\") pod \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " Jan 26 17:15:09 crc kubenswrapper[4856]: I0126 17:15:09.679075 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-build-ca-bundles\") pod \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " Jan 26 17:15:09 crc kubenswrapper[4856]: I0126 17:15:09.679134 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-container-storage-run\") pod \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " Jan 26 17:15:09 crc kubenswrapper[4856]: I0126 17:15:09.679154 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-builder-dockercfg-8h4xs-pull\") pod \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " Jan 26 17:15:09 crc kubenswrapper[4856]: I0126 17:15:09.679261 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-builder-dockercfg-8h4xs-push\") pod \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " Jan 26 17:15:09 crc kubenswrapper[4856]: I0126 17:15:09.679287 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-node-pullsecrets\") pod \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\" (UID: \"e3f6dcf4-c152-4a81-8e1d-1fdf469be581\") " Jan 26 17:15:09 crc kubenswrapper[4856]: I0126 17:15:09.679751 4856 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-buildcachedir\") on node \"crc\" DevicePath \"\"" Jan 26 17:15:09 crc kubenswrapper[4856]: I0126 17:15:09.679796 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "e3f6dcf4-c152-4a81-8e1d-1fdf469be581" (UID: "e3f6dcf4-c152-4a81-8e1d-1fdf469be581"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:15:09 crc kubenswrapper[4856]: I0126 17:15:09.680494 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "e3f6dcf4-c152-4a81-8e1d-1fdf469be581" (UID: "e3f6dcf4-c152-4a81-8e1d-1fdf469be581"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:15:09 crc kubenswrapper[4856]: I0126 17:15:09.681142 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "e3f6dcf4-c152-4a81-8e1d-1fdf469be581" (UID: "e3f6dcf4-c152-4a81-8e1d-1fdf469be581"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:15:09 crc kubenswrapper[4856]: I0126 17:15:09.681889 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "e3f6dcf4-c152-4a81-8e1d-1fdf469be581" (UID: "e3f6dcf4-c152-4a81-8e1d-1fdf469be581"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:15:09 crc kubenswrapper[4856]: I0126 17:15:09.680463 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "e3f6dcf4-c152-4a81-8e1d-1fdf469be581" (UID: "e3f6dcf4-c152-4a81-8e1d-1fdf469be581"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:15:09 crc kubenswrapper[4856]: I0126 17:15:09.686261 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-builder-dockercfg-8h4xs-pull" (OuterVolumeSpecName: "builder-dockercfg-8h4xs-pull") pod "e3f6dcf4-c152-4a81-8e1d-1fdf469be581" (UID: "e3f6dcf4-c152-4a81-8e1d-1fdf469be581"). InnerVolumeSpecName "builder-dockercfg-8h4xs-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:15:09 crc kubenswrapper[4856]: I0126 17:15:09.686382 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-kube-api-access-spxgc" (OuterVolumeSpecName: "kube-api-access-spxgc") pod "e3f6dcf4-c152-4a81-8e1d-1fdf469be581" (UID: "e3f6dcf4-c152-4a81-8e1d-1fdf469be581"). InnerVolumeSpecName "kube-api-access-spxgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:15:09 crc kubenswrapper[4856]: I0126 17:15:09.688656 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-builder-dockercfg-8h4xs-push" (OuterVolumeSpecName: "builder-dockercfg-8h4xs-push") pod "e3f6dcf4-c152-4a81-8e1d-1fdf469be581" (UID: "e3f6dcf4-c152-4a81-8e1d-1fdf469be581"). InnerVolumeSpecName "builder-dockercfg-8h4xs-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:15:09 crc kubenswrapper[4856]: I0126 17:15:09.727309 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "e3f6dcf4-c152-4a81-8e1d-1fdf469be581" (UID: "e3f6dcf4-c152-4a81-8e1d-1fdf469be581"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:15:09 crc kubenswrapper[4856]: I0126 17:15:09.781007 4856 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-container-storage-run\") on node \"crc\" DevicePath \"\"" Jan 26 17:15:09 crc kubenswrapper[4856]: I0126 17:15:09.781043 4856 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-builder-dockercfg-8h4xs-pull\") on node \"crc\" DevicePath \"\"" Jan 26 17:15:09 crc kubenswrapper[4856]: I0126 17:15:09.781055 4856 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-builder-dockercfg-8h4xs-push\") on node \"crc\" DevicePath \"\"" Jan 26 17:15:09 crc kubenswrapper[4856]: I0126 17:15:09.781067 4856 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Jan 26 17:15:09 crc kubenswrapper[4856]: I0126 17:15:09.781079 4856 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 26 17:15:09 crc kubenswrapper[4856]: I0126 17:15:09.781089 4856 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-build-system-configs\") on node \"crc\" DevicePath \"\"" Jan 26 17:15:09 crc kubenswrapper[4856]: I0126 17:15:09.781102 4856 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-buildworkdir\") on node \"crc\" DevicePath \"\"" Jan 26 17:15:09 crc kubenswrapper[4856]: I0126 17:15:09.781113 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-spxgc\" (UniqueName: \"kubernetes.io/projected/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-kube-api-access-spxgc\") on node \"crc\" DevicePath \"\"" Jan 26 17:15:09 crc kubenswrapper[4856]: I0126 17:15:09.781124 4856 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 26 17:15:09 crc kubenswrapper[4856]: I0126 17:15:09.856618 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "e3f6dcf4-c152-4a81-8e1d-1fdf469be581" (UID: "e3f6dcf4-c152-4a81-8e1d-1fdf469be581"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:15:09 crc kubenswrapper[4856]: I0126 17:15:09.881928 4856 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-build-blob-cache\") on node \"crc\" DevicePath \"\"" Jan 26 17:15:10 crc kubenswrapper[4856]: I0126 17:15:10.341799 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"e3f6dcf4-c152-4a81-8e1d-1fdf469be581","Type":"ContainerDied","Data":"f38fe849edfe940888da0c7e9589bf8433e33392b1573a13a0d673b63831ce2b"} Jan 26 17:15:10 crc kubenswrapper[4856]: I0126 17:15:10.341873 4856 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f38fe849edfe940888da0c7e9589bf8433e33392b1573a13a0d673b63831ce2b" Jan 26 17:15:10 crc kubenswrapper[4856]: I0126 17:15:10.341893 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Jan 26 17:15:11 crc kubenswrapper[4856]: I0126 17:15:11.666790 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "e3f6dcf4-c152-4a81-8e1d-1fdf469be581" (UID: "e3f6dcf4-c152-4a81-8e1d-1fdf469be581"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:15:11 crc kubenswrapper[4856]: I0126 17:15:11.688437 4856 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/e3f6dcf4-c152-4a81-8e1d-1fdf469be581-container-storage-root\") on node \"crc\" DevicePath \"\"" Jan 26 17:15:14 crc kubenswrapper[4856]: I0126 17:15:14.820728 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/smart-gateway-operator-1-build"] Jan 26 17:15:14 crc kubenswrapper[4856]: E0126 17:15:14.821387 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3f6dcf4-c152-4a81-8e1d-1fdf469be581" containerName="git-clone" Jan 26 17:15:14 crc kubenswrapper[4856]: I0126 17:15:14.821410 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3f6dcf4-c152-4a81-8e1d-1fdf469be581" containerName="git-clone" Jan 26 17:15:14 crc kubenswrapper[4856]: E0126 17:15:14.821423 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3f61a40-5427-4fe8-89d2-92b71f9e1052" containerName="collect-profiles" Jan 26 17:15:14 crc kubenswrapper[4856]: I0126 17:15:14.821429 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3f61a40-5427-4fe8-89d2-92b71f9e1052" containerName="collect-profiles" Jan 26 17:15:14 crc kubenswrapper[4856]: E0126 17:15:14.821441 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3f6dcf4-c152-4a81-8e1d-1fdf469be581" containerName="manage-dockerfile" Jan 26 17:15:14 crc kubenswrapper[4856]: I0126 17:15:14.821447 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3f6dcf4-c152-4a81-8e1d-1fdf469be581" containerName="manage-dockerfile" Jan 26 17:15:14 crc kubenswrapper[4856]: E0126 17:15:14.821456 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3f6dcf4-c152-4a81-8e1d-1fdf469be581" containerName="docker-build" Jan 26 17:15:14 crc kubenswrapper[4856]: I0126 17:15:14.821462 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3f6dcf4-c152-4a81-8e1d-1fdf469be581" containerName="docker-build" Jan 26 17:15:14 crc kubenswrapper[4856]: I0126 17:15:14.821626 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3f6dcf4-c152-4a81-8e1d-1fdf469be581" containerName="docker-build" Jan 26 17:15:14 crc kubenswrapper[4856]: I0126 17:15:14.821639 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3f61a40-5427-4fe8-89d2-92b71f9e1052" containerName="collect-profiles" Jan 26 17:15:14 crc kubenswrapper[4856]: I0126 17:15:14.822728 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-1-build" Jan 26 17:15:14 crc kubenswrapper[4856]: I0126 17:15:14.825087 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-1-sys-config" Jan 26 17:15:14 crc kubenswrapper[4856]: I0126 17:15:14.825811 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-1-global-ca" Jan 26 17:15:14 crc kubenswrapper[4856]: I0126 17:15:14.826054 4856 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-8h4xs" Jan 26 17:15:14 crc kubenswrapper[4856]: I0126 17:15:14.826115 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-1-ca" Jan 26 17:15:14 crc kubenswrapper[4856]: I0126 17:15:14.837418 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-1-build"] Jan 26 17:15:14 crc kubenswrapper[4856]: I0126 17:15:14.955350 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5ed22249-c992-4d01-a0ec-110a1ff4f786-build-proxy-ca-bundles\") pod \"smart-gateway-operator-1-build\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 26 17:15:14 crc kubenswrapper[4856]: I0126 17:15:14.955421 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/5ed22249-c992-4d01-a0ec-110a1ff4f786-builder-dockercfg-8h4xs-pull\") pod \"smart-gateway-operator-1-build\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 26 17:15:14 crc kubenswrapper[4856]: I0126 17:15:14.955502 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/5ed22249-c992-4d01-a0ec-110a1ff4f786-build-system-configs\") pod \"smart-gateway-operator-1-build\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 26 17:15:14 crc kubenswrapper[4856]: I0126 17:15:14.955541 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/5ed22249-c992-4d01-a0ec-110a1ff4f786-buildcachedir\") pod \"smart-gateway-operator-1-build\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 26 17:15:14 crc kubenswrapper[4856]: I0126 17:15:14.955559 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/5ed22249-c992-4d01-a0ec-110a1ff4f786-buildworkdir\") pod \"smart-gateway-operator-1-build\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 26 17:15:14 crc kubenswrapper[4856]: I0126 17:15:14.955574 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/5ed22249-c992-4d01-a0ec-110a1ff4f786-container-storage-root\") pod \"smart-gateway-operator-1-build\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 26 17:15:14 crc kubenswrapper[4856]: I0126 17:15:14.955594 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7q4r\" (UniqueName: \"kubernetes.io/projected/5ed22249-c992-4d01-a0ec-110a1ff4f786-kube-api-access-d7q4r\") pod \"smart-gateway-operator-1-build\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 26 17:15:14 crc kubenswrapper[4856]: I0126 17:15:14.955626 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5ed22249-c992-4d01-a0ec-110a1ff4f786-node-pullsecrets\") pod \"smart-gateway-operator-1-build\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 26 17:15:14 crc kubenswrapper[4856]: I0126 17:15:14.955853 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/5ed22249-c992-4d01-a0ec-110a1ff4f786-builder-dockercfg-8h4xs-push\") pod \"smart-gateway-operator-1-build\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 26 17:15:14 crc kubenswrapper[4856]: I0126 17:15:14.955972 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5ed22249-c992-4d01-a0ec-110a1ff4f786-build-ca-bundles\") pod \"smart-gateway-operator-1-build\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 26 17:15:14 crc kubenswrapper[4856]: I0126 17:15:14.956182 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/5ed22249-c992-4d01-a0ec-110a1ff4f786-build-blob-cache\") pod \"smart-gateway-operator-1-build\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 26 17:15:14 crc kubenswrapper[4856]: I0126 17:15:14.956226 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/5ed22249-c992-4d01-a0ec-110a1ff4f786-container-storage-run\") pod \"smart-gateway-operator-1-build\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 26 17:15:15 crc kubenswrapper[4856]: I0126 17:15:15.057396 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/5ed22249-c992-4d01-a0ec-110a1ff4f786-buildcachedir\") pod \"smart-gateway-operator-1-build\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 26 17:15:15 crc kubenswrapper[4856]: I0126 17:15:15.057469 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/5ed22249-c992-4d01-a0ec-110a1ff4f786-buildworkdir\") pod \"smart-gateway-operator-1-build\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 26 17:15:15 crc kubenswrapper[4856]: I0126 17:15:15.057503 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/5ed22249-c992-4d01-a0ec-110a1ff4f786-container-storage-root\") pod \"smart-gateway-operator-1-build\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 26 17:15:15 crc kubenswrapper[4856]: I0126 17:15:15.057568 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7q4r\" (UniqueName: \"kubernetes.io/projected/5ed22249-c992-4d01-a0ec-110a1ff4f786-kube-api-access-d7q4r\") pod \"smart-gateway-operator-1-build\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 26 17:15:15 crc kubenswrapper[4856]: I0126 17:15:15.057584 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/5ed22249-c992-4d01-a0ec-110a1ff4f786-buildcachedir\") pod \"smart-gateway-operator-1-build\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 26 17:15:15 crc kubenswrapper[4856]: I0126 17:15:15.057614 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5ed22249-c992-4d01-a0ec-110a1ff4f786-node-pullsecrets\") pod \"smart-gateway-operator-1-build\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 26 17:15:15 crc kubenswrapper[4856]: I0126 17:15:15.057730 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/5ed22249-c992-4d01-a0ec-110a1ff4f786-builder-dockercfg-8h4xs-push\") pod \"smart-gateway-operator-1-build\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 26 17:15:15 crc kubenswrapper[4856]: I0126 17:15:15.057851 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5ed22249-c992-4d01-a0ec-110a1ff4f786-node-pullsecrets\") pod \"smart-gateway-operator-1-build\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 26 17:15:15 crc kubenswrapper[4856]: I0126 17:15:15.058315 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/5ed22249-c992-4d01-a0ec-110a1ff4f786-buildworkdir\") pod \"smart-gateway-operator-1-build\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 26 17:15:15 crc kubenswrapper[4856]: I0126 17:15:15.058357 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/5ed22249-c992-4d01-a0ec-110a1ff4f786-container-storage-root\") pod \"smart-gateway-operator-1-build\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 26 17:15:15 crc kubenswrapper[4856]: I0126 17:15:15.058923 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5ed22249-c992-4d01-a0ec-110a1ff4f786-build-ca-bundles\") pod \"smart-gateway-operator-1-build\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 26 17:15:15 crc kubenswrapper[4856]: I0126 17:15:15.059031 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/5ed22249-c992-4d01-a0ec-110a1ff4f786-build-blob-cache\") pod \"smart-gateway-operator-1-build\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 26 17:15:15 crc kubenswrapper[4856]: I0126 17:15:15.059076 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/5ed22249-c992-4d01-a0ec-110a1ff4f786-container-storage-run\") pod \"smart-gateway-operator-1-build\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 26 17:15:15 crc kubenswrapper[4856]: I0126 17:15:15.059133 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5ed22249-c992-4d01-a0ec-110a1ff4f786-build-proxy-ca-bundles\") pod \"smart-gateway-operator-1-build\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 26 17:15:15 crc kubenswrapper[4856]: I0126 17:15:15.059176 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/5ed22249-c992-4d01-a0ec-110a1ff4f786-builder-dockercfg-8h4xs-pull\") pod \"smart-gateway-operator-1-build\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 26 17:15:15 crc kubenswrapper[4856]: I0126 17:15:15.059214 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/5ed22249-c992-4d01-a0ec-110a1ff4f786-build-system-configs\") pod \"smart-gateway-operator-1-build\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 26 17:15:15 crc kubenswrapper[4856]: I0126 17:15:15.059437 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/5ed22249-c992-4d01-a0ec-110a1ff4f786-build-blob-cache\") pod \"smart-gateway-operator-1-build\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 26 17:15:15 crc kubenswrapper[4856]: I0126 17:15:15.059788 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/5ed22249-c992-4d01-a0ec-110a1ff4f786-container-storage-run\") pod \"smart-gateway-operator-1-build\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 26 17:15:15 crc kubenswrapper[4856]: I0126 17:15:15.060441 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/5ed22249-c992-4d01-a0ec-110a1ff4f786-build-system-configs\") pod \"smart-gateway-operator-1-build\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 26 17:15:15 crc kubenswrapper[4856]: I0126 17:15:15.060827 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5ed22249-c992-4d01-a0ec-110a1ff4f786-build-proxy-ca-bundles\") pod \"smart-gateway-operator-1-build\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 26 17:15:15 crc kubenswrapper[4856]: I0126 17:15:15.063898 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/5ed22249-c992-4d01-a0ec-110a1ff4f786-builder-dockercfg-8h4xs-pull\") pod \"smart-gateway-operator-1-build\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 26 17:15:15 crc kubenswrapper[4856]: I0126 17:15:15.064367 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5ed22249-c992-4d01-a0ec-110a1ff4f786-build-ca-bundles\") pod \"smart-gateway-operator-1-build\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 26 17:15:15 crc kubenswrapper[4856]: I0126 17:15:15.065032 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/5ed22249-c992-4d01-a0ec-110a1ff4f786-builder-dockercfg-8h4xs-push\") pod \"smart-gateway-operator-1-build\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 26 17:15:15 crc kubenswrapper[4856]: I0126 17:15:15.081064 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7q4r\" (UniqueName: \"kubernetes.io/projected/5ed22249-c992-4d01-a0ec-110a1ff4f786-kube-api-access-d7q4r\") pod \"smart-gateway-operator-1-build\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " pod="service-telemetry/smart-gateway-operator-1-build" Jan 26 17:15:15 crc kubenswrapper[4856]: I0126 17:15:15.143398 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-1-build" Jan 26 17:15:15 crc kubenswrapper[4856]: I0126 17:15:15.582335 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-1-build"] Jan 26 17:15:16 crc kubenswrapper[4856]: I0126 17:15:16.416792 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-1-build" event={"ID":"5ed22249-c992-4d01-a0ec-110a1ff4f786","Type":"ContainerStarted","Data":"bbaa0e65ef3404876abbdc6bdb86225f8808634f7616aeafb5182074b30443dd"} Jan 26 17:15:16 crc kubenswrapper[4856]: I0126 17:15:16.417128 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-1-build" event={"ID":"5ed22249-c992-4d01-a0ec-110a1ff4f786","Type":"ContainerStarted","Data":"32c29441f42e73b7bc00f264639641dd4797b6d25e6fdbaa53648b8d228dc4d6"} Jan 26 17:15:17 crc kubenswrapper[4856]: I0126 17:15:17.426059 4856 generic.go:334] "Generic (PLEG): container finished" podID="5ed22249-c992-4d01-a0ec-110a1ff4f786" containerID="bbaa0e65ef3404876abbdc6bdb86225f8808634f7616aeafb5182074b30443dd" exitCode=0 Jan 26 17:15:17 crc kubenswrapper[4856]: I0126 17:15:17.426167 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-1-build" event={"ID":"5ed22249-c992-4d01-a0ec-110a1ff4f786","Type":"ContainerDied","Data":"bbaa0e65ef3404876abbdc6bdb86225f8808634f7616aeafb5182074b30443dd"} Jan 26 17:15:18 crc kubenswrapper[4856]: I0126 17:15:18.438863 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-1-build" event={"ID":"5ed22249-c992-4d01-a0ec-110a1ff4f786","Type":"ContainerStarted","Data":"fefa4e9ff30a3471c605b939c41d47c1a8dfc5a17b0199c647a6644529b4b943"} Jan 26 17:15:18 crc kubenswrapper[4856]: I0126 17:15:18.462330 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/smart-gateway-operator-1-build" podStartSLOduration=4.462293644 podStartE2EDuration="4.462293644s" podCreationTimestamp="2026-01-26 17:15:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:15:18.460606986 +0000 UTC m=+1014.413861007" watchObservedRunningTime="2026-01-26 17:15:18.462293644 +0000 UTC m=+1014.415547675" Jan 26 17:15:25 crc kubenswrapper[4856]: I0126 17:15:25.384923 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/smart-gateway-operator-1-build"] Jan 26 17:15:25 crc kubenswrapper[4856]: I0126 17:15:25.385688 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/smart-gateway-operator-1-build" podUID="5ed22249-c992-4d01-a0ec-110a1ff4f786" containerName="docker-build" containerID="cri-o://fefa4e9ff30a3471c605b939c41d47c1a8dfc5a17b0199c647a6644529b4b943" gracePeriod=30 Jan 26 17:15:26 crc kubenswrapper[4856]: I0126 17:15:26.938603 4856 patch_prober.go:28] interesting pod/machine-config-daemon-xm9cq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:15:26 crc kubenswrapper[4856]: I0126 17:15:26.939102 4856 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:15:27 crc kubenswrapper[4856]: I0126 17:15:27.111419 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/smart-gateway-operator-2-build"] Jan 26 17:15:27 crc kubenswrapper[4856]: I0126 17:15:27.113027 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-2-build" Jan 26 17:15:27 crc kubenswrapper[4856]: I0126 17:15:27.115014 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-2-sys-config" Jan 26 17:15:27 crc kubenswrapper[4856]: I0126 17:15:27.115087 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-2-global-ca" Jan 26 17:15:27 crc kubenswrapper[4856]: I0126 17:15:27.115706 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-2-ca" Jan 26 17:15:27 crc kubenswrapper[4856]: I0126 17:15:27.133243 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-2-build"] Jan 26 17:15:27 crc kubenswrapper[4856]: I0126 17:15:27.292081 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a2268784-2c29-45fa-8bbc-4426f4c566b6-node-pullsecrets\") pod \"smart-gateway-operator-2-build\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 26 17:15:27 crc kubenswrapper[4856]: I0126 17:15:27.292141 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/a2268784-2c29-45fa-8bbc-4426f4c566b6-buildworkdir\") pod \"smart-gateway-operator-2-build\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 26 17:15:27 crc kubenswrapper[4856]: I0126 17:15:27.292172 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/a2268784-2c29-45fa-8bbc-4426f4c566b6-container-storage-root\") pod \"smart-gateway-operator-2-build\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 26 17:15:27 crc kubenswrapper[4856]: I0126 17:15:27.292227 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a2268784-2c29-45fa-8bbc-4426f4c566b6-build-proxy-ca-bundles\") pod \"smart-gateway-operator-2-build\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 26 17:15:27 crc kubenswrapper[4856]: I0126 17:15:27.292850 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/a2268784-2c29-45fa-8bbc-4426f4c566b6-container-storage-run\") pod \"smart-gateway-operator-2-build\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 26 17:15:27 crc kubenswrapper[4856]: I0126 17:15:27.293010 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/a2268784-2c29-45fa-8bbc-4426f4c566b6-build-system-configs\") pod \"smart-gateway-operator-2-build\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 26 17:15:27 crc kubenswrapper[4856]: I0126 17:15:27.293118 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/a2268784-2c29-45fa-8bbc-4426f4c566b6-builder-dockercfg-8h4xs-pull\") pod \"smart-gateway-operator-2-build\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 26 17:15:27 crc kubenswrapper[4856]: I0126 17:15:27.293170 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a2268784-2c29-45fa-8bbc-4426f4c566b6-build-ca-bundles\") pod \"smart-gateway-operator-2-build\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 26 17:15:27 crc kubenswrapper[4856]: I0126 17:15:27.293237 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hs6s\" (UniqueName: \"kubernetes.io/projected/a2268784-2c29-45fa-8bbc-4426f4c566b6-kube-api-access-7hs6s\") pod \"smart-gateway-operator-2-build\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 26 17:15:27 crc kubenswrapper[4856]: I0126 17:15:27.293280 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/a2268784-2c29-45fa-8bbc-4426f4c566b6-buildcachedir\") pod \"smart-gateway-operator-2-build\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 26 17:15:27 crc kubenswrapper[4856]: I0126 17:15:27.293363 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/a2268784-2c29-45fa-8bbc-4426f4c566b6-builder-dockercfg-8h4xs-push\") pod \"smart-gateway-operator-2-build\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 26 17:15:27 crc kubenswrapper[4856]: I0126 17:15:27.293443 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/a2268784-2c29-45fa-8bbc-4426f4c566b6-build-blob-cache\") pod \"smart-gateway-operator-2-build\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 26 17:15:27 crc kubenswrapper[4856]: I0126 17:15:27.394951 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/a2268784-2c29-45fa-8bbc-4426f4c566b6-container-storage-run\") pod \"smart-gateway-operator-2-build\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 26 17:15:27 crc kubenswrapper[4856]: I0126 17:15:27.395062 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/a2268784-2c29-45fa-8bbc-4426f4c566b6-build-system-configs\") pod \"smart-gateway-operator-2-build\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 26 17:15:27 crc kubenswrapper[4856]: I0126 17:15:27.395127 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/a2268784-2c29-45fa-8bbc-4426f4c566b6-builder-dockercfg-8h4xs-pull\") pod \"smart-gateway-operator-2-build\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 26 17:15:27 crc kubenswrapper[4856]: I0126 17:15:27.395174 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a2268784-2c29-45fa-8bbc-4426f4c566b6-build-ca-bundles\") pod \"smart-gateway-operator-2-build\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 26 17:15:27 crc kubenswrapper[4856]: I0126 17:15:27.395229 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hs6s\" (UniqueName: \"kubernetes.io/projected/a2268784-2c29-45fa-8bbc-4426f4c566b6-kube-api-access-7hs6s\") pod \"smart-gateway-operator-2-build\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 26 17:15:27 crc kubenswrapper[4856]: I0126 17:15:27.395269 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/a2268784-2c29-45fa-8bbc-4426f4c566b6-buildcachedir\") pod \"smart-gateway-operator-2-build\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 26 17:15:27 crc kubenswrapper[4856]: I0126 17:15:27.395320 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/a2268784-2c29-45fa-8bbc-4426f4c566b6-builder-dockercfg-8h4xs-push\") pod \"smart-gateway-operator-2-build\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 26 17:15:27 crc kubenswrapper[4856]: I0126 17:15:27.395394 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/a2268784-2c29-45fa-8bbc-4426f4c566b6-build-blob-cache\") pod \"smart-gateway-operator-2-build\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 26 17:15:27 crc kubenswrapper[4856]: I0126 17:15:27.395469 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a2268784-2c29-45fa-8bbc-4426f4c566b6-node-pullsecrets\") pod \"smart-gateway-operator-2-build\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 26 17:15:27 crc kubenswrapper[4856]: I0126 17:15:27.395474 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/a2268784-2c29-45fa-8bbc-4426f4c566b6-container-storage-run\") pod \"smart-gateway-operator-2-build\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 26 17:15:27 crc kubenswrapper[4856]: I0126 17:15:27.395506 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/a2268784-2c29-45fa-8bbc-4426f4c566b6-buildworkdir\") pod \"smart-gateway-operator-2-build\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 26 17:15:27 crc kubenswrapper[4856]: I0126 17:15:27.395601 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/a2268784-2c29-45fa-8bbc-4426f4c566b6-container-storage-root\") pod \"smart-gateway-operator-2-build\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 26 17:15:27 crc kubenswrapper[4856]: I0126 17:15:27.395640 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a2268784-2c29-45fa-8bbc-4426f4c566b6-build-proxy-ca-bundles\") pod \"smart-gateway-operator-2-build\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 26 17:15:27 crc kubenswrapper[4856]: I0126 17:15:27.395795 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/a2268784-2c29-45fa-8bbc-4426f4c566b6-build-blob-cache\") pod \"smart-gateway-operator-2-build\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 26 17:15:27 crc kubenswrapper[4856]: I0126 17:15:27.395472 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/a2268784-2c29-45fa-8bbc-4426f4c566b6-buildcachedir\") pod \"smart-gateway-operator-2-build\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 26 17:15:27 crc kubenswrapper[4856]: I0126 17:15:27.396447 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/a2268784-2c29-45fa-8bbc-4426f4c566b6-build-system-configs\") pod \"smart-gateway-operator-2-build\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 26 17:15:27 crc kubenswrapper[4856]: I0126 17:15:27.396512 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/a2268784-2c29-45fa-8bbc-4426f4c566b6-buildworkdir\") pod \"smart-gateway-operator-2-build\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 26 17:15:27 crc kubenswrapper[4856]: I0126 17:15:27.396843 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/a2268784-2c29-45fa-8bbc-4426f4c566b6-container-storage-root\") pod \"smart-gateway-operator-2-build\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 26 17:15:27 crc kubenswrapper[4856]: I0126 17:15:27.396947 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a2268784-2c29-45fa-8bbc-4426f4c566b6-build-ca-bundles\") pod \"smart-gateway-operator-2-build\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 26 17:15:27 crc kubenswrapper[4856]: I0126 17:15:27.396942 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a2268784-2c29-45fa-8bbc-4426f4c566b6-node-pullsecrets\") pod \"smart-gateway-operator-2-build\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 26 17:15:27 crc kubenswrapper[4856]: I0126 17:15:27.398243 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a2268784-2c29-45fa-8bbc-4426f4c566b6-build-proxy-ca-bundles\") pod \"smart-gateway-operator-2-build\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 26 17:15:27 crc kubenswrapper[4856]: I0126 17:15:27.405395 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/a2268784-2c29-45fa-8bbc-4426f4c566b6-builder-dockercfg-8h4xs-pull\") pod \"smart-gateway-operator-2-build\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 26 17:15:27 crc kubenswrapper[4856]: I0126 17:15:27.412163 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/a2268784-2c29-45fa-8bbc-4426f4c566b6-builder-dockercfg-8h4xs-push\") pod \"smart-gateway-operator-2-build\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 26 17:15:27 crc kubenswrapper[4856]: I0126 17:15:27.436948 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hs6s\" (UniqueName: \"kubernetes.io/projected/a2268784-2c29-45fa-8bbc-4426f4c566b6-kube-api-access-7hs6s\") pod \"smart-gateway-operator-2-build\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " pod="service-telemetry/smart-gateway-operator-2-build" Jan 26 17:15:27 crc kubenswrapper[4856]: I0126 17:15:27.726969 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-2-build" Jan 26 17:15:28 crc kubenswrapper[4856]: I0126 17:15:28.016986 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-2-build"] Jan 26 17:15:28 crc kubenswrapper[4856]: I0126 17:15:28.512346 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"a2268784-2c29-45fa-8bbc-4426f4c566b6","Type":"ContainerStarted","Data":"06f2ce65083300356c56ba4f8a7f06492d9d84894e9cf8a9d78cb9fbe7bdcb6c"} Jan 26 17:15:29 crc kubenswrapper[4856]: I0126 17:15:29.523518 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-1-build_5ed22249-c992-4d01-a0ec-110a1ff4f786/docker-build/0.log" Jan 26 17:15:29 crc kubenswrapper[4856]: I0126 17:15:29.523975 4856 generic.go:334] "Generic (PLEG): container finished" podID="5ed22249-c992-4d01-a0ec-110a1ff4f786" containerID="fefa4e9ff30a3471c605b939c41d47c1a8dfc5a17b0199c647a6644529b4b943" exitCode=1 Jan 26 17:15:29 crc kubenswrapper[4856]: I0126 17:15:29.524018 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-1-build" event={"ID":"5ed22249-c992-4d01-a0ec-110a1ff4f786","Type":"ContainerDied","Data":"fefa4e9ff30a3471c605b939c41d47c1a8dfc5a17b0199c647a6644529b4b943"} Jan 26 17:15:29 crc kubenswrapper[4856]: I0126 17:15:29.524807 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"a2268784-2c29-45fa-8bbc-4426f4c566b6","Type":"ContainerStarted","Data":"f1c91d37a26cde48c83a1293d17ffc8b332fdc2b1d02b20144bd258945a7c047"} Jan 26 17:15:29 crc kubenswrapper[4856]: I0126 17:15:29.699499 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-1-build_5ed22249-c992-4d01-a0ec-110a1ff4f786/docker-build/0.log" Jan 26 17:15:29 crc kubenswrapper[4856]: I0126 17:15:29.699907 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-1-build" Jan 26 17:15:29 crc kubenswrapper[4856]: I0126 17:15:29.928740 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5ed22249-c992-4d01-a0ec-110a1ff4f786-build-ca-bundles\") pod \"5ed22249-c992-4d01-a0ec-110a1ff4f786\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " Jan 26 17:15:29 crc kubenswrapper[4856]: I0126 17:15:29.928787 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/5ed22249-c992-4d01-a0ec-110a1ff4f786-build-system-configs\") pod \"5ed22249-c992-4d01-a0ec-110a1ff4f786\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " Jan 26 17:15:29 crc kubenswrapper[4856]: I0126 17:15:29.928819 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/5ed22249-c992-4d01-a0ec-110a1ff4f786-builder-dockercfg-8h4xs-pull\") pod \"5ed22249-c992-4d01-a0ec-110a1ff4f786\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " Jan 26 17:15:29 crc kubenswrapper[4856]: I0126 17:15:29.928868 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/5ed22249-c992-4d01-a0ec-110a1ff4f786-builder-dockercfg-8h4xs-push\") pod \"5ed22249-c992-4d01-a0ec-110a1ff4f786\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " Jan 26 17:15:29 crc kubenswrapper[4856]: I0126 17:15:29.928911 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/5ed22249-c992-4d01-a0ec-110a1ff4f786-buildworkdir\") pod \"5ed22249-c992-4d01-a0ec-110a1ff4f786\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " Jan 26 17:15:29 crc kubenswrapper[4856]: I0126 17:15:29.928980 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5ed22249-c992-4d01-a0ec-110a1ff4f786-node-pullsecrets\") pod \"5ed22249-c992-4d01-a0ec-110a1ff4f786\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " Jan 26 17:15:29 crc kubenswrapper[4856]: I0126 17:15:29.929011 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5ed22249-c992-4d01-a0ec-110a1ff4f786-build-proxy-ca-bundles\") pod \"5ed22249-c992-4d01-a0ec-110a1ff4f786\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " Jan 26 17:15:29 crc kubenswrapper[4856]: I0126 17:15:29.929058 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/5ed22249-c992-4d01-a0ec-110a1ff4f786-build-blob-cache\") pod \"5ed22249-c992-4d01-a0ec-110a1ff4f786\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " Jan 26 17:15:29 crc kubenswrapper[4856]: I0126 17:15:29.929087 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d7q4r\" (UniqueName: \"kubernetes.io/projected/5ed22249-c992-4d01-a0ec-110a1ff4f786-kube-api-access-d7q4r\") pod \"5ed22249-c992-4d01-a0ec-110a1ff4f786\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " Jan 26 17:15:29 crc kubenswrapper[4856]: I0126 17:15:29.929131 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/5ed22249-c992-4d01-a0ec-110a1ff4f786-buildcachedir\") pod \"5ed22249-c992-4d01-a0ec-110a1ff4f786\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " Jan 26 17:15:29 crc kubenswrapper[4856]: I0126 17:15:29.929126 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5ed22249-c992-4d01-a0ec-110a1ff4f786-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "5ed22249-c992-4d01-a0ec-110a1ff4f786" (UID: "5ed22249-c992-4d01-a0ec-110a1ff4f786"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:15:29 crc kubenswrapper[4856]: I0126 17:15:29.929159 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/5ed22249-c992-4d01-a0ec-110a1ff4f786-container-storage-run\") pod \"5ed22249-c992-4d01-a0ec-110a1ff4f786\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " Jan 26 17:15:29 crc kubenswrapper[4856]: I0126 17:15:29.929183 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/5ed22249-c992-4d01-a0ec-110a1ff4f786-container-storage-root\") pod \"5ed22249-c992-4d01-a0ec-110a1ff4f786\" (UID: \"5ed22249-c992-4d01-a0ec-110a1ff4f786\") " Jan 26 17:15:29 crc kubenswrapper[4856]: I0126 17:15:29.929443 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ed22249-c992-4d01-a0ec-110a1ff4f786-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "5ed22249-c992-4d01-a0ec-110a1ff4f786" (UID: "5ed22249-c992-4d01-a0ec-110a1ff4f786"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:15:29 crc kubenswrapper[4856]: I0126 17:15:29.929461 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5ed22249-c992-4d01-a0ec-110a1ff4f786-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "5ed22249-c992-4d01-a0ec-110a1ff4f786" (UID: "5ed22249-c992-4d01-a0ec-110a1ff4f786"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:15:29 crc kubenswrapper[4856]: I0126 17:15:29.929481 4856 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5ed22249-c992-4d01-a0ec-110a1ff4f786-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Jan 26 17:15:29 crc kubenswrapper[4856]: I0126 17:15:29.929723 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ed22249-c992-4d01-a0ec-110a1ff4f786-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "5ed22249-c992-4d01-a0ec-110a1ff4f786" (UID: "5ed22249-c992-4d01-a0ec-110a1ff4f786"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:15:29 crc kubenswrapper[4856]: I0126 17:15:29.929756 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ed22249-c992-4d01-a0ec-110a1ff4f786-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "5ed22249-c992-4d01-a0ec-110a1ff4f786" (UID: "5ed22249-c992-4d01-a0ec-110a1ff4f786"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:15:29 crc kubenswrapper[4856]: I0126 17:15:29.930040 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ed22249-c992-4d01-a0ec-110a1ff4f786-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "5ed22249-c992-4d01-a0ec-110a1ff4f786" (UID: "5ed22249-c992-4d01-a0ec-110a1ff4f786"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:15:29 crc kubenswrapper[4856]: I0126 17:15:29.931825 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ed22249-c992-4d01-a0ec-110a1ff4f786-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "5ed22249-c992-4d01-a0ec-110a1ff4f786" (UID: "5ed22249-c992-4d01-a0ec-110a1ff4f786"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:15:29 crc kubenswrapper[4856]: I0126 17:15:29.935154 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ed22249-c992-4d01-a0ec-110a1ff4f786-kube-api-access-d7q4r" (OuterVolumeSpecName: "kube-api-access-d7q4r") pod "5ed22249-c992-4d01-a0ec-110a1ff4f786" (UID: "5ed22249-c992-4d01-a0ec-110a1ff4f786"). InnerVolumeSpecName "kube-api-access-d7q4r". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:15:29 crc kubenswrapper[4856]: I0126 17:15:29.952513 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ed22249-c992-4d01-a0ec-110a1ff4f786-builder-dockercfg-8h4xs-push" (OuterVolumeSpecName: "builder-dockercfg-8h4xs-push") pod "5ed22249-c992-4d01-a0ec-110a1ff4f786" (UID: "5ed22249-c992-4d01-a0ec-110a1ff4f786"). InnerVolumeSpecName "builder-dockercfg-8h4xs-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:15:29 crc kubenswrapper[4856]: I0126 17:15:29.952647 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ed22249-c992-4d01-a0ec-110a1ff4f786-builder-dockercfg-8h4xs-pull" (OuterVolumeSpecName: "builder-dockercfg-8h4xs-pull") pod "5ed22249-c992-4d01-a0ec-110a1ff4f786" (UID: "5ed22249-c992-4d01-a0ec-110a1ff4f786"). InnerVolumeSpecName "builder-dockercfg-8h4xs-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:15:30 crc kubenswrapper[4856]: I0126 17:15:30.031201 4856 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/5ed22249-c992-4d01-a0ec-110a1ff4f786-container-storage-run\") on node \"crc\" DevicePath \"\"" Jan 26 17:15:30 crc kubenswrapper[4856]: I0126 17:15:30.031253 4856 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5ed22249-c992-4d01-a0ec-110a1ff4f786-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 26 17:15:30 crc kubenswrapper[4856]: I0126 17:15:30.031272 4856 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/5ed22249-c992-4d01-a0ec-110a1ff4f786-build-system-configs\") on node \"crc\" DevicePath \"\"" Jan 26 17:15:30 crc kubenswrapper[4856]: I0126 17:15:30.031290 4856 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/5ed22249-c992-4d01-a0ec-110a1ff4f786-builder-dockercfg-8h4xs-pull\") on node \"crc\" DevicePath \"\"" Jan 26 17:15:30 crc kubenswrapper[4856]: I0126 17:15:30.031308 4856 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/5ed22249-c992-4d01-a0ec-110a1ff4f786-builder-dockercfg-8h4xs-push\") on node \"crc\" DevicePath \"\"" Jan 26 17:15:30 crc kubenswrapper[4856]: I0126 17:15:30.031326 4856 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/5ed22249-c992-4d01-a0ec-110a1ff4f786-buildworkdir\") on node \"crc\" DevicePath \"\"" Jan 26 17:15:30 crc kubenswrapper[4856]: I0126 17:15:30.031344 4856 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5ed22249-c992-4d01-a0ec-110a1ff4f786-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 26 17:15:30 crc kubenswrapper[4856]: I0126 17:15:30.031361 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d7q4r\" (UniqueName: \"kubernetes.io/projected/5ed22249-c992-4d01-a0ec-110a1ff4f786-kube-api-access-d7q4r\") on node \"crc\" DevicePath \"\"" Jan 26 17:15:30 crc kubenswrapper[4856]: I0126 17:15:30.031378 4856 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/5ed22249-c992-4d01-a0ec-110a1ff4f786-buildcachedir\") on node \"crc\" DevicePath \"\"" Jan 26 17:15:30 crc kubenswrapper[4856]: I0126 17:15:30.323879 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ed22249-c992-4d01-a0ec-110a1ff4f786-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "5ed22249-c992-4d01-a0ec-110a1ff4f786" (UID: "5ed22249-c992-4d01-a0ec-110a1ff4f786"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:15:30 crc kubenswrapper[4856]: I0126 17:15:30.335356 4856 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/5ed22249-c992-4d01-a0ec-110a1ff4f786-container-storage-root\") on node \"crc\" DevicePath \"\"" Jan 26 17:15:30 crc kubenswrapper[4856]: I0126 17:15:30.534676 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-1-build_5ed22249-c992-4d01-a0ec-110a1ff4f786/docker-build/0.log" Jan 26 17:15:30 crc kubenswrapper[4856]: I0126 17:15:30.536933 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-1-build" Jan 26 17:15:30 crc kubenswrapper[4856]: I0126 17:15:30.536943 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-1-build" event={"ID":"5ed22249-c992-4d01-a0ec-110a1ff4f786","Type":"ContainerDied","Data":"32c29441f42e73b7bc00f264639641dd4797b6d25e6fdbaa53648b8d228dc4d6"} Jan 26 17:15:30 crc kubenswrapper[4856]: I0126 17:15:30.537224 4856 scope.go:117] "RemoveContainer" containerID="fefa4e9ff30a3471c605b939c41d47c1a8dfc5a17b0199c647a6644529b4b943" Jan 26 17:15:32 crc kubenswrapper[4856]: I0126 17:15:32.941983 4856 scope.go:117] "RemoveContainer" containerID="bbaa0e65ef3404876abbdc6bdb86225f8808634f7616aeafb5182074b30443dd" Jan 26 17:15:33 crc kubenswrapper[4856]: I0126 17:15:33.161380 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ed22249-c992-4d01-a0ec-110a1ff4f786-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "5ed22249-c992-4d01-a0ec-110a1ff4f786" (UID: "5ed22249-c992-4d01-a0ec-110a1ff4f786"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:15:33 crc kubenswrapper[4856]: I0126 17:15:33.174715 4856 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/5ed22249-c992-4d01-a0ec-110a1ff4f786-build-blob-cache\") on node \"crc\" DevicePath \"\"" Jan 26 17:15:33 crc kubenswrapper[4856]: I0126 17:15:33.277094 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/smart-gateway-operator-1-build"] Jan 26 17:15:33 crc kubenswrapper[4856]: I0126 17:15:33.289485 4856 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/smart-gateway-operator-1-build"] Jan 26 17:15:33 crc kubenswrapper[4856]: I0126 17:15:33.405429 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ed22249-c992-4d01-a0ec-110a1ff4f786" path="/var/lib/kubelet/pods/5ed22249-c992-4d01-a0ec-110a1ff4f786/volumes" Jan 26 17:15:35 crc kubenswrapper[4856]: I0126 17:15:35.578361 4856 generic.go:334] "Generic (PLEG): container finished" podID="a2268784-2c29-45fa-8bbc-4426f4c566b6" containerID="f1c91d37a26cde48c83a1293d17ffc8b332fdc2b1d02b20144bd258945a7c047" exitCode=0 Jan 26 17:15:35 crc kubenswrapper[4856]: I0126 17:15:35.578477 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"a2268784-2c29-45fa-8bbc-4426f4c566b6","Type":"ContainerDied","Data":"f1c91d37a26cde48c83a1293d17ffc8b332fdc2b1d02b20144bd258945a7c047"} Jan 26 17:15:36 crc kubenswrapper[4856]: I0126 17:15:36.591086 4856 generic.go:334] "Generic (PLEG): container finished" podID="a2268784-2c29-45fa-8bbc-4426f4c566b6" containerID="78d8d20e1d5fb801d43742459729ca3548bb6acf3b8d3566e927a42c56e4febf" exitCode=0 Jan 26 17:15:36 crc kubenswrapper[4856]: I0126 17:15:36.591237 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"a2268784-2c29-45fa-8bbc-4426f4c566b6","Type":"ContainerDied","Data":"78d8d20e1d5fb801d43742459729ca3548bb6acf3b8d3566e927a42c56e4febf"} Jan 26 17:15:36 crc kubenswrapper[4856]: I0126 17:15:36.640847 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-2-build_a2268784-2c29-45fa-8bbc-4426f4c566b6/manage-dockerfile/0.log" Jan 26 17:15:37 crc kubenswrapper[4856]: I0126 17:15:37.607179 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"a2268784-2c29-45fa-8bbc-4426f4c566b6","Type":"ContainerStarted","Data":"bace1809b288c35ceefa0274d2e72a4e0c2a1f126e377b5ce2ae65e151a5a665"} Jan 26 17:15:37 crc kubenswrapper[4856]: I0126 17:15:37.641829 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/smart-gateway-operator-2-build" podStartSLOduration=10.641802646 podStartE2EDuration="10.641802646s" podCreationTimestamp="2026-01-26 17:15:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:15:37.639903182 +0000 UTC m=+1033.593157163" watchObservedRunningTime="2026-01-26 17:15:37.641802646 +0000 UTC m=+1033.595056627" Jan 26 17:15:56 crc kubenswrapper[4856]: I0126 17:15:56.939447 4856 patch_prober.go:28] interesting pod/machine-config-daemon-xm9cq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:15:56 crc kubenswrapper[4856]: I0126 17:15:56.940153 4856 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:16:19 crc kubenswrapper[4856]: I0126 17:16:19.354122 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bz275"] Jan 26 17:16:19 crc kubenswrapper[4856]: E0126 17:16:19.356022 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ed22249-c992-4d01-a0ec-110a1ff4f786" containerName="docker-build" Jan 26 17:16:19 crc kubenswrapper[4856]: I0126 17:16:19.356165 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ed22249-c992-4d01-a0ec-110a1ff4f786" containerName="docker-build" Jan 26 17:16:19 crc kubenswrapper[4856]: E0126 17:16:19.356192 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ed22249-c992-4d01-a0ec-110a1ff4f786" containerName="manage-dockerfile" Jan 26 17:16:19 crc kubenswrapper[4856]: I0126 17:16:19.356201 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ed22249-c992-4d01-a0ec-110a1ff4f786" containerName="manage-dockerfile" Jan 26 17:16:19 crc kubenswrapper[4856]: I0126 17:16:19.357417 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ed22249-c992-4d01-a0ec-110a1ff4f786" containerName="docker-build" Jan 26 17:16:19 crc kubenswrapper[4856]: I0126 17:16:19.361241 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bz275"] Jan 26 17:16:19 crc kubenswrapper[4856]: I0126 17:16:19.362608 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bz275" Jan 26 17:16:19 crc kubenswrapper[4856]: I0126 17:16:19.504830 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqldd\" (UniqueName: \"kubernetes.io/projected/4dfaabcd-3733-443e-8a38-060ad2953eec-kube-api-access-tqldd\") pod \"certified-operators-bz275\" (UID: \"4dfaabcd-3733-443e-8a38-060ad2953eec\") " pod="openshift-marketplace/certified-operators-bz275" Jan 26 17:16:19 crc kubenswrapper[4856]: I0126 17:16:19.504909 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4dfaabcd-3733-443e-8a38-060ad2953eec-utilities\") pod \"certified-operators-bz275\" (UID: \"4dfaabcd-3733-443e-8a38-060ad2953eec\") " pod="openshift-marketplace/certified-operators-bz275" Jan 26 17:16:19 crc kubenswrapper[4856]: I0126 17:16:19.504987 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4dfaabcd-3733-443e-8a38-060ad2953eec-catalog-content\") pod \"certified-operators-bz275\" (UID: \"4dfaabcd-3733-443e-8a38-060ad2953eec\") " pod="openshift-marketplace/certified-operators-bz275" Jan 26 17:16:19 crc kubenswrapper[4856]: I0126 17:16:19.606595 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqldd\" (UniqueName: \"kubernetes.io/projected/4dfaabcd-3733-443e-8a38-060ad2953eec-kube-api-access-tqldd\") pod \"certified-operators-bz275\" (UID: \"4dfaabcd-3733-443e-8a38-060ad2953eec\") " pod="openshift-marketplace/certified-operators-bz275" Jan 26 17:16:19 crc kubenswrapper[4856]: I0126 17:16:19.606695 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4dfaabcd-3733-443e-8a38-060ad2953eec-utilities\") pod \"certified-operators-bz275\" (UID: \"4dfaabcd-3733-443e-8a38-060ad2953eec\") " pod="openshift-marketplace/certified-operators-bz275" Jan 26 17:16:19 crc kubenswrapper[4856]: I0126 17:16:19.606747 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4dfaabcd-3733-443e-8a38-060ad2953eec-catalog-content\") pod \"certified-operators-bz275\" (UID: \"4dfaabcd-3733-443e-8a38-060ad2953eec\") " pod="openshift-marketplace/certified-operators-bz275" Jan 26 17:16:19 crc kubenswrapper[4856]: I0126 17:16:19.607412 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4dfaabcd-3733-443e-8a38-060ad2953eec-catalog-content\") pod \"certified-operators-bz275\" (UID: \"4dfaabcd-3733-443e-8a38-060ad2953eec\") " pod="openshift-marketplace/certified-operators-bz275" Jan 26 17:16:19 crc kubenswrapper[4856]: I0126 17:16:19.607407 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4dfaabcd-3733-443e-8a38-060ad2953eec-utilities\") pod \"certified-operators-bz275\" (UID: \"4dfaabcd-3733-443e-8a38-060ad2953eec\") " pod="openshift-marketplace/certified-operators-bz275" Jan 26 17:16:19 crc kubenswrapper[4856]: I0126 17:16:19.631079 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqldd\" (UniqueName: \"kubernetes.io/projected/4dfaabcd-3733-443e-8a38-060ad2953eec-kube-api-access-tqldd\") pod \"certified-operators-bz275\" (UID: \"4dfaabcd-3733-443e-8a38-060ad2953eec\") " pod="openshift-marketplace/certified-operators-bz275" Jan 26 17:16:19 crc kubenswrapper[4856]: I0126 17:16:19.684515 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bz275" Jan 26 17:16:20 crc kubenswrapper[4856]: I0126 17:16:20.011199 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bz275"] Jan 26 17:16:20 crc kubenswrapper[4856]: I0126 17:16:20.349068 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bz275" event={"ID":"4dfaabcd-3733-443e-8a38-060ad2953eec","Type":"ContainerStarted","Data":"90b10b1ad76d4b052f858ba03b4cb3df4bcd4b4057edd943836d778d0526b08f"} Jan 26 17:16:24 crc kubenswrapper[4856]: I0126 17:16:24.390168 4856 generic.go:334] "Generic (PLEG): container finished" podID="4dfaabcd-3733-443e-8a38-060ad2953eec" containerID="ea9dd6ca56fab3dd3d36f63cf819d952d1b23acac4807908175e381c3bc121d4" exitCode=0 Jan 26 17:16:24 crc kubenswrapper[4856]: I0126 17:16:24.390217 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bz275" event={"ID":"4dfaabcd-3733-443e-8a38-060ad2953eec","Type":"ContainerDied","Data":"ea9dd6ca56fab3dd3d36f63cf819d952d1b23acac4807908175e381c3bc121d4"} Jan 26 17:16:24 crc kubenswrapper[4856]: I0126 17:16:24.392313 4856 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 26 17:16:26 crc kubenswrapper[4856]: I0126 17:16:26.406359 4856 generic.go:334] "Generic (PLEG): container finished" podID="4dfaabcd-3733-443e-8a38-060ad2953eec" containerID="4a6f6ebaf882577a384aa190207606f891d548d42a3382b291a7ded23f0cec89" exitCode=0 Jan 26 17:16:26 crc kubenswrapper[4856]: I0126 17:16:26.406538 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bz275" event={"ID":"4dfaabcd-3733-443e-8a38-060ad2953eec","Type":"ContainerDied","Data":"4a6f6ebaf882577a384aa190207606f891d548d42a3382b291a7ded23f0cec89"} Jan 26 17:16:26 crc kubenswrapper[4856]: I0126 17:16:26.939648 4856 patch_prober.go:28] interesting pod/machine-config-daemon-xm9cq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:16:26 crc kubenswrapper[4856]: I0126 17:16:26.939850 4856 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:16:26 crc kubenswrapper[4856]: I0126 17:16:26.939983 4856 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" Jan 26 17:16:26 crc kubenswrapper[4856]: I0126 17:16:26.941340 4856 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fdaad4602089daad40b0395fbc761e615a8ba2a94c8f5b977142a787034cddb7"} pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 26 17:16:26 crc kubenswrapper[4856]: I0126 17:16:26.941476 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" containerName="machine-config-daemon" containerID="cri-o://fdaad4602089daad40b0395fbc761e615a8ba2a94c8f5b977142a787034cddb7" gracePeriod=600 Jan 26 17:16:27 crc kubenswrapper[4856]: I0126 17:16:27.415159 4856 generic.go:334] "Generic (PLEG): container finished" podID="63c75ede-5170-4db0-811b-5217ef8d72b3" containerID="fdaad4602089daad40b0395fbc761e615a8ba2a94c8f5b977142a787034cddb7" exitCode=0 Jan 26 17:16:27 crc kubenswrapper[4856]: I0126 17:16:27.415270 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" event={"ID":"63c75ede-5170-4db0-811b-5217ef8d72b3","Type":"ContainerDied","Data":"fdaad4602089daad40b0395fbc761e615a8ba2a94c8f5b977142a787034cddb7"} Jan 26 17:16:27 crc kubenswrapper[4856]: I0126 17:16:27.415666 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" event={"ID":"63c75ede-5170-4db0-811b-5217ef8d72b3","Type":"ContainerStarted","Data":"5846ab4d870be5fcbab796c3e27690d2c13d129480d6fcd21b3b0d1c535f0cff"} Jan 26 17:16:27 crc kubenswrapper[4856]: I0126 17:16:27.415708 4856 scope.go:117] "RemoveContainer" containerID="bb3fb578d0ea2b4eb264b402043faa4d1923f5d38749a2ee2c65b084c2e291bd" Jan 26 17:16:27 crc kubenswrapper[4856]: I0126 17:16:27.418348 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bz275" event={"ID":"4dfaabcd-3733-443e-8a38-060ad2953eec","Type":"ContainerStarted","Data":"6b1c37d94c31c69c26080700f0bce3069c7c6f0ae57f4ffc06c5518d82c24f5f"} Jan 26 17:16:27 crc kubenswrapper[4856]: I0126 17:16:27.451098 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bz275" podStartSLOduration=5.764174881 podStartE2EDuration="8.451052735s" podCreationTimestamp="2026-01-26 17:16:19 +0000 UTC" firstStartedPulling="2026-01-26 17:16:24.391940264 +0000 UTC m=+1080.345194245" lastFinishedPulling="2026-01-26 17:16:27.078818108 +0000 UTC m=+1083.032072099" observedRunningTime="2026-01-26 17:16:27.449376447 +0000 UTC m=+1083.402630438" watchObservedRunningTime="2026-01-26 17:16:27.451052735 +0000 UTC m=+1083.404306716" Jan 26 17:16:29 crc kubenswrapper[4856]: I0126 17:16:29.684816 4856 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bz275" Jan 26 17:16:29 crc kubenswrapper[4856]: I0126 17:16:29.685500 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bz275" Jan 26 17:16:29 crc kubenswrapper[4856]: I0126 17:16:29.736224 4856 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bz275" Jan 26 17:16:39 crc kubenswrapper[4856]: I0126 17:16:39.733600 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bz275" Jan 26 17:16:40 crc kubenswrapper[4856]: I0126 17:16:40.693620 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bz275"] Jan 26 17:16:40 crc kubenswrapper[4856]: I0126 17:16:40.694102 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bz275" podUID="4dfaabcd-3733-443e-8a38-060ad2953eec" containerName="registry-server" containerID="cri-o://6b1c37d94c31c69c26080700f0bce3069c7c6f0ae57f4ffc06c5518d82c24f5f" gracePeriod=2 Jan 26 17:16:43 crc kubenswrapper[4856]: I0126 17:16:43.554686 4856 generic.go:334] "Generic (PLEG): container finished" podID="4dfaabcd-3733-443e-8a38-060ad2953eec" containerID="6b1c37d94c31c69c26080700f0bce3069c7c6f0ae57f4ffc06c5518d82c24f5f" exitCode=0 Jan 26 17:16:43 crc kubenswrapper[4856]: I0126 17:16:43.554771 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bz275" event={"ID":"4dfaabcd-3733-443e-8a38-060ad2953eec","Type":"ContainerDied","Data":"6b1c37d94c31c69c26080700f0bce3069c7c6f0ae57f4ffc06c5518d82c24f5f"} Jan 26 17:16:43 crc kubenswrapper[4856]: I0126 17:16:43.805909 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bz275" Jan 26 17:16:43 crc kubenswrapper[4856]: I0126 17:16:43.972734 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4dfaabcd-3733-443e-8a38-060ad2953eec-catalog-content\") pod \"4dfaabcd-3733-443e-8a38-060ad2953eec\" (UID: \"4dfaabcd-3733-443e-8a38-060ad2953eec\") " Jan 26 17:16:43 crc kubenswrapper[4856]: I0126 17:16:43.973035 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tqldd\" (UniqueName: \"kubernetes.io/projected/4dfaabcd-3733-443e-8a38-060ad2953eec-kube-api-access-tqldd\") pod \"4dfaabcd-3733-443e-8a38-060ad2953eec\" (UID: \"4dfaabcd-3733-443e-8a38-060ad2953eec\") " Jan 26 17:16:43 crc kubenswrapper[4856]: I0126 17:16:43.973136 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4dfaabcd-3733-443e-8a38-060ad2953eec-utilities\") pod \"4dfaabcd-3733-443e-8a38-060ad2953eec\" (UID: \"4dfaabcd-3733-443e-8a38-060ad2953eec\") " Jan 26 17:16:43 crc kubenswrapper[4856]: I0126 17:16:43.974293 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4dfaabcd-3733-443e-8a38-060ad2953eec-utilities" (OuterVolumeSpecName: "utilities") pod "4dfaabcd-3733-443e-8a38-060ad2953eec" (UID: "4dfaabcd-3733-443e-8a38-060ad2953eec"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:16:43 crc kubenswrapper[4856]: I0126 17:16:43.980259 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4dfaabcd-3733-443e-8a38-060ad2953eec-kube-api-access-tqldd" (OuterVolumeSpecName: "kube-api-access-tqldd") pod "4dfaabcd-3733-443e-8a38-060ad2953eec" (UID: "4dfaabcd-3733-443e-8a38-060ad2953eec"). InnerVolumeSpecName "kube-api-access-tqldd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:16:44 crc kubenswrapper[4856]: I0126 17:16:44.021267 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4dfaabcd-3733-443e-8a38-060ad2953eec-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4dfaabcd-3733-443e-8a38-060ad2953eec" (UID: "4dfaabcd-3733-443e-8a38-060ad2953eec"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:16:44 crc kubenswrapper[4856]: I0126 17:16:44.074844 4856 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4dfaabcd-3733-443e-8a38-060ad2953eec-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 17:16:44 crc kubenswrapper[4856]: I0126 17:16:44.074892 4856 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4dfaabcd-3733-443e-8a38-060ad2953eec-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 17:16:44 crc kubenswrapper[4856]: I0126 17:16:44.074908 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqldd\" (UniqueName: \"kubernetes.io/projected/4dfaabcd-3733-443e-8a38-060ad2953eec-kube-api-access-tqldd\") on node \"crc\" DevicePath \"\"" Jan 26 17:16:44 crc kubenswrapper[4856]: I0126 17:16:44.564163 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bz275" event={"ID":"4dfaabcd-3733-443e-8a38-060ad2953eec","Type":"ContainerDied","Data":"90b10b1ad76d4b052f858ba03b4cb3df4bcd4b4057edd943836d778d0526b08f"} Jan 26 17:16:44 crc kubenswrapper[4856]: I0126 17:16:44.564426 4856 scope.go:117] "RemoveContainer" containerID="6b1c37d94c31c69c26080700f0bce3069c7c6f0ae57f4ffc06c5518d82c24f5f" Jan 26 17:16:44 crc kubenswrapper[4856]: I0126 17:16:44.564246 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bz275" Jan 26 17:16:44 crc kubenswrapper[4856]: I0126 17:16:44.584987 4856 scope.go:117] "RemoveContainer" containerID="4a6f6ebaf882577a384aa190207606f891d548d42a3382b291a7ded23f0cec89" Jan 26 17:16:44 crc kubenswrapper[4856]: I0126 17:16:44.594631 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bz275"] Jan 26 17:16:44 crc kubenswrapper[4856]: I0126 17:16:44.607373 4856 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bz275"] Jan 26 17:16:44 crc kubenswrapper[4856]: I0126 17:16:44.616119 4856 scope.go:117] "RemoveContainer" containerID="ea9dd6ca56fab3dd3d36f63cf819d952d1b23acac4807908175e381c3bc121d4" Jan 26 17:16:45 crc kubenswrapper[4856]: I0126 17:16:45.404002 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4dfaabcd-3733-443e-8a38-060ad2953eec" path="/var/lib/kubelet/pods/4dfaabcd-3733-443e-8a38-060ad2953eec/volumes" Jan 26 17:17:02 crc kubenswrapper[4856]: I0126 17:17:02.720485 4856 generic.go:334] "Generic (PLEG): container finished" podID="a2268784-2c29-45fa-8bbc-4426f4c566b6" containerID="bace1809b288c35ceefa0274d2e72a4e0c2a1f126e377b5ce2ae65e151a5a665" exitCode=0 Jan 26 17:17:02 crc kubenswrapper[4856]: I0126 17:17:02.720835 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"a2268784-2c29-45fa-8bbc-4426f4c566b6","Type":"ContainerDied","Data":"bace1809b288c35ceefa0274d2e72a4e0c2a1f126e377b5ce2ae65e151a5a665"} Jan 26 17:17:04 crc kubenswrapper[4856]: I0126 17:17:04.043279 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-2-build" Jan 26 17:17:04 crc kubenswrapper[4856]: I0126 17:17:04.175855 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/a2268784-2c29-45fa-8bbc-4426f4c566b6-container-storage-root\") pod \"a2268784-2c29-45fa-8bbc-4426f4c566b6\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " Jan 26 17:17:04 crc kubenswrapper[4856]: I0126 17:17:04.175961 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/a2268784-2c29-45fa-8bbc-4426f4c566b6-buildcachedir\") pod \"a2268784-2c29-45fa-8bbc-4426f4c566b6\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " Jan 26 17:17:04 crc kubenswrapper[4856]: I0126 17:17:04.176009 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/a2268784-2c29-45fa-8bbc-4426f4c566b6-build-blob-cache\") pod \"a2268784-2c29-45fa-8bbc-4426f4c566b6\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " Jan 26 17:17:04 crc kubenswrapper[4856]: I0126 17:17:04.176051 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a2268784-2c29-45fa-8bbc-4426f4c566b6-build-ca-bundles\") pod \"a2268784-2c29-45fa-8bbc-4426f4c566b6\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " Jan 26 17:17:04 crc kubenswrapper[4856]: I0126 17:17:04.176074 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a2268784-2c29-45fa-8bbc-4426f4c566b6-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "a2268784-2c29-45fa-8bbc-4426f4c566b6" (UID: "a2268784-2c29-45fa-8bbc-4426f4c566b6"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:17:04 crc kubenswrapper[4856]: I0126 17:17:04.176080 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a2268784-2c29-45fa-8bbc-4426f4c566b6-node-pullsecrets\") pod \"a2268784-2c29-45fa-8bbc-4426f4c566b6\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " Jan 26 17:17:04 crc kubenswrapper[4856]: I0126 17:17:04.176105 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a2268784-2c29-45fa-8bbc-4426f4c566b6-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "a2268784-2c29-45fa-8bbc-4426f4c566b6" (UID: "a2268784-2c29-45fa-8bbc-4426f4c566b6"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:17:04 crc kubenswrapper[4856]: I0126 17:17:04.176137 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/a2268784-2c29-45fa-8bbc-4426f4c566b6-builder-dockercfg-8h4xs-pull\") pod \"a2268784-2c29-45fa-8bbc-4426f4c566b6\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " Jan 26 17:17:04 crc kubenswrapper[4856]: I0126 17:17:04.176163 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a2268784-2c29-45fa-8bbc-4426f4c566b6-build-proxy-ca-bundles\") pod \"a2268784-2c29-45fa-8bbc-4426f4c566b6\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " Jan 26 17:17:04 crc kubenswrapper[4856]: I0126 17:17:04.176196 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/a2268784-2c29-45fa-8bbc-4426f4c566b6-container-storage-run\") pod \"a2268784-2c29-45fa-8bbc-4426f4c566b6\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " Jan 26 17:17:04 crc kubenswrapper[4856]: I0126 17:17:04.176232 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/a2268784-2c29-45fa-8bbc-4426f4c566b6-builder-dockercfg-8h4xs-push\") pod \"a2268784-2c29-45fa-8bbc-4426f4c566b6\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " Jan 26 17:17:04 crc kubenswrapper[4856]: I0126 17:17:04.176254 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/a2268784-2c29-45fa-8bbc-4426f4c566b6-build-system-configs\") pod \"a2268784-2c29-45fa-8bbc-4426f4c566b6\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " Jan 26 17:17:04 crc kubenswrapper[4856]: I0126 17:17:04.176275 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7hs6s\" (UniqueName: \"kubernetes.io/projected/a2268784-2c29-45fa-8bbc-4426f4c566b6-kube-api-access-7hs6s\") pod \"a2268784-2c29-45fa-8bbc-4426f4c566b6\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " Jan 26 17:17:04 crc kubenswrapper[4856]: I0126 17:17:04.176310 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/a2268784-2c29-45fa-8bbc-4426f4c566b6-buildworkdir\") pod \"a2268784-2c29-45fa-8bbc-4426f4c566b6\" (UID: \"a2268784-2c29-45fa-8bbc-4426f4c566b6\") " Jan 26 17:17:04 crc kubenswrapper[4856]: I0126 17:17:04.176623 4856 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/a2268784-2c29-45fa-8bbc-4426f4c566b6-buildcachedir\") on node \"crc\" DevicePath \"\"" Jan 26 17:17:04 crc kubenswrapper[4856]: I0126 17:17:04.176644 4856 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a2268784-2c29-45fa-8bbc-4426f4c566b6-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Jan 26 17:17:04 crc kubenswrapper[4856]: I0126 17:17:04.177388 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2268784-2c29-45fa-8bbc-4426f4c566b6-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "a2268784-2c29-45fa-8bbc-4426f4c566b6" (UID: "a2268784-2c29-45fa-8bbc-4426f4c566b6"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:17:04 crc kubenswrapper[4856]: I0126 17:17:04.177781 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2268784-2c29-45fa-8bbc-4426f4c566b6-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "a2268784-2c29-45fa-8bbc-4426f4c566b6" (UID: "a2268784-2c29-45fa-8bbc-4426f4c566b6"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:17:04 crc kubenswrapper[4856]: I0126 17:17:04.177839 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2268784-2c29-45fa-8bbc-4426f4c566b6-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "a2268784-2c29-45fa-8bbc-4426f4c566b6" (UID: "a2268784-2c29-45fa-8bbc-4426f4c566b6"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:17:04 crc kubenswrapper[4856]: I0126 17:17:04.178727 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2268784-2c29-45fa-8bbc-4426f4c566b6-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "a2268784-2c29-45fa-8bbc-4426f4c566b6" (UID: "a2268784-2c29-45fa-8bbc-4426f4c566b6"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:17:04 crc kubenswrapper[4856]: I0126 17:17:04.182246 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2268784-2c29-45fa-8bbc-4426f4c566b6-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "a2268784-2c29-45fa-8bbc-4426f4c566b6" (UID: "a2268784-2c29-45fa-8bbc-4426f4c566b6"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:17:04 crc kubenswrapper[4856]: I0126 17:17:04.183783 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2268784-2c29-45fa-8bbc-4426f4c566b6-kube-api-access-7hs6s" (OuterVolumeSpecName: "kube-api-access-7hs6s") pod "a2268784-2c29-45fa-8bbc-4426f4c566b6" (UID: "a2268784-2c29-45fa-8bbc-4426f4c566b6"). InnerVolumeSpecName "kube-api-access-7hs6s". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:17:04 crc kubenswrapper[4856]: I0126 17:17:04.184243 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2268784-2c29-45fa-8bbc-4426f4c566b6-builder-dockercfg-8h4xs-push" (OuterVolumeSpecName: "builder-dockercfg-8h4xs-push") pod "a2268784-2c29-45fa-8bbc-4426f4c566b6" (UID: "a2268784-2c29-45fa-8bbc-4426f4c566b6"). InnerVolumeSpecName "builder-dockercfg-8h4xs-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:17:04 crc kubenswrapper[4856]: I0126 17:17:04.185948 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2268784-2c29-45fa-8bbc-4426f4c566b6-builder-dockercfg-8h4xs-pull" (OuterVolumeSpecName: "builder-dockercfg-8h4xs-pull") pod "a2268784-2c29-45fa-8bbc-4426f4c566b6" (UID: "a2268784-2c29-45fa-8bbc-4426f4c566b6"). InnerVolumeSpecName "builder-dockercfg-8h4xs-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:17:04 crc kubenswrapper[4856]: I0126 17:17:04.278396 4856 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a2268784-2c29-45fa-8bbc-4426f4c566b6-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 26 17:17:04 crc kubenswrapper[4856]: I0126 17:17:04.278434 4856 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/a2268784-2c29-45fa-8bbc-4426f4c566b6-builder-dockercfg-8h4xs-pull\") on node \"crc\" DevicePath \"\"" Jan 26 17:17:04 crc kubenswrapper[4856]: I0126 17:17:04.278448 4856 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a2268784-2c29-45fa-8bbc-4426f4c566b6-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 26 17:17:04 crc kubenswrapper[4856]: I0126 17:17:04.278458 4856 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/a2268784-2c29-45fa-8bbc-4426f4c566b6-container-storage-run\") on node \"crc\" DevicePath \"\"" Jan 26 17:17:04 crc kubenswrapper[4856]: I0126 17:17:04.278470 4856 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/a2268784-2c29-45fa-8bbc-4426f4c566b6-builder-dockercfg-8h4xs-push\") on node \"crc\" DevicePath \"\"" Jan 26 17:17:04 crc kubenswrapper[4856]: I0126 17:17:04.278480 4856 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/a2268784-2c29-45fa-8bbc-4426f4c566b6-build-system-configs\") on node \"crc\" DevicePath \"\"" Jan 26 17:17:04 crc kubenswrapper[4856]: I0126 17:17:04.278489 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7hs6s\" (UniqueName: \"kubernetes.io/projected/a2268784-2c29-45fa-8bbc-4426f4c566b6-kube-api-access-7hs6s\") on node \"crc\" DevicePath \"\"" Jan 26 17:17:04 crc kubenswrapper[4856]: I0126 17:17:04.278499 4856 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/a2268784-2c29-45fa-8bbc-4426f4c566b6-buildworkdir\") on node \"crc\" DevicePath \"\"" Jan 26 17:17:04 crc kubenswrapper[4856]: I0126 17:17:04.416831 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2268784-2c29-45fa-8bbc-4426f4c566b6-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "a2268784-2c29-45fa-8bbc-4426f4c566b6" (UID: "a2268784-2c29-45fa-8bbc-4426f4c566b6"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:17:04 crc kubenswrapper[4856]: I0126 17:17:04.482018 4856 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/a2268784-2c29-45fa-8bbc-4426f4c566b6-build-blob-cache\") on node \"crc\" DevicePath \"\"" Jan 26 17:17:04 crc kubenswrapper[4856]: I0126 17:17:04.735544 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"a2268784-2c29-45fa-8bbc-4426f4c566b6","Type":"ContainerDied","Data":"06f2ce65083300356c56ba4f8a7f06492d9d84894e9cf8a9d78cb9fbe7bdcb6c"} Jan 26 17:17:04 crc kubenswrapper[4856]: I0126 17:17:04.735608 4856 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="06f2ce65083300356c56ba4f8a7f06492d9d84894e9cf8a9d78cb9fbe7bdcb6c" Jan 26 17:17:04 crc kubenswrapper[4856]: I0126 17:17:04.735650 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-2-build" Jan 26 17:17:06 crc kubenswrapper[4856]: I0126 17:17:06.236577 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2268784-2c29-45fa-8bbc-4426f4c566b6-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "a2268784-2c29-45fa-8bbc-4426f4c566b6" (UID: "a2268784-2c29-45fa-8bbc-4426f4c566b6"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:17:06 crc kubenswrapper[4856]: I0126 17:17:06.309232 4856 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/a2268784-2c29-45fa-8bbc-4426f4c566b6-container-storage-root\") on node \"crc\" DevicePath \"\"" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.003315 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/sg-core-1-build"] Jan 26 17:17:09 crc kubenswrapper[4856]: E0126 17:17:09.004076 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dfaabcd-3733-443e-8a38-060ad2953eec" containerName="extract-content" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.004097 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dfaabcd-3733-443e-8a38-060ad2953eec" containerName="extract-content" Jan 26 17:17:09 crc kubenswrapper[4856]: E0126 17:17:09.004115 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2268784-2c29-45fa-8bbc-4426f4c566b6" containerName="git-clone" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.004125 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2268784-2c29-45fa-8bbc-4426f4c566b6" containerName="git-clone" Jan 26 17:17:09 crc kubenswrapper[4856]: E0126 17:17:09.004142 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dfaabcd-3733-443e-8a38-060ad2953eec" containerName="registry-server" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.004152 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dfaabcd-3733-443e-8a38-060ad2953eec" containerName="registry-server" Jan 26 17:17:09 crc kubenswrapper[4856]: E0126 17:17:09.004169 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dfaabcd-3733-443e-8a38-060ad2953eec" containerName="extract-utilities" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.004179 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dfaabcd-3733-443e-8a38-060ad2953eec" containerName="extract-utilities" Jan 26 17:17:09 crc kubenswrapper[4856]: E0126 17:17:09.004201 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2268784-2c29-45fa-8bbc-4426f4c566b6" containerName="docker-build" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.004211 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2268784-2c29-45fa-8bbc-4426f4c566b6" containerName="docker-build" Jan 26 17:17:09 crc kubenswrapper[4856]: E0126 17:17:09.004227 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2268784-2c29-45fa-8bbc-4426f4c566b6" containerName="manage-dockerfile" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.004237 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2268784-2c29-45fa-8bbc-4426f4c566b6" containerName="manage-dockerfile" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.004398 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="4dfaabcd-3733-443e-8a38-060ad2953eec" containerName="registry-server" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.004423 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2268784-2c29-45fa-8bbc-4426f4c566b6" containerName="docker-build" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.005404 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-1-build" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.008872 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-core-1-ca" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.008888 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-core-1-sys-config" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.009236 4856 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-8h4xs" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.011709 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-core-1-global-ca" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.029436 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-core-1-build"] Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.154654 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/845832d9-625c-452e-b900-4e3c2df2ef4d-build-blob-cache\") pod \"sg-core-1-build\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " pod="service-telemetry/sg-core-1-build" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.155195 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/845832d9-625c-452e-b900-4e3c2df2ef4d-buildcachedir\") pod \"sg-core-1-build\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " pod="service-telemetry/sg-core-1-build" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.155335 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/845832d9-625c-452e-b900-4e3c2df2ef4d-container-storage-root\") pod \"sg-core-1-build\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " pod="service-telemetry/sg-core-1-build" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.155588 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/845832d9-625c-452e-b900-4e3c2df2ef4d-node-pullsecrets\") pod \"sg-core-1-build\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " pod="service-telemetry/sg-core-1-build" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.155832 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/845832d9-625c-452e-b900-4e3c2df2ef4d-build-proxy-ca-bundles\") pod \"sg-core-1-build\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " pod="service-telemetry/sg-core-1-build" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.155974 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/845832d9-625c-452e-b900-4e3c2df2ef4d-builder-dockercfg-8h4xs-push\") pod \"sg-core-1-build\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " pod="service-telemetry/sg-core-1-build" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.156031 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/845832d9-625c-452e-b900-4e3c2df2ef4d-build-ca-bundles\") pod \"sg-core-1-build\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " pod="service-telemetry/sg-core-1-build" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.156049 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/845832d9-625c-452e-b900-4e3c2df2ef4d-container-storage-run\") pod \"sg-core-1-build\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " pod="service-telemetry/sg-core-1-build" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.156082 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/845832d9-625c-452e-b900-4e3c2df2ef4d-buildworkdir\") pod \"sg-core-1-build\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " pod="service-telemetry/sg-core-1-build" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.156100 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/845832d9-625c-452e-b900-4e3c2df2ef4d-builder-dockercfg-8h4xs-pull\") pod \"sg-core-1-build\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " pod="service-telemetry/sg-core-1-build" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.156341 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/845832d9-625c-452e-b900-4e3c2df2ef4d-build-system-configs\") pod \"sg-core-1-build\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " pod="service-telemetry/sg-core-1-build" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.156381 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7glkd\" (UniqueName: \"kubernetes.io/projected/845832d9-625c-452e-b900-4e3c2df2ef4d-kube-api-access-7glkd\") pod \"sg-core-1-build\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " pod="service-telemetry/sg-core-1-build" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.258045 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/845832d9-625c-452e-b900-4e3c2df2ef4d-build-system-configs\") pod \"sg-core-1-build\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " pod="service-telemetry/sg-core-1-build" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.258089 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7glkd\" (UniqueName: \"kubernetes.io/projected/845832d9-625c-452e-b900-4e3c2df2ef4d-kube-api-access-7glkd\") pod \"sg-core-1-build\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " pod="service-telemetry/sg-core-1-build" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.258130 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/845832d9-625c-452e-b900-4e3c2df2ef4d-build-blob-cache\") pod \"sg-core-1-build\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " pod="service-telemetry/sg-core-1-build" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.258146 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/845832d9-625c-452e-b900-4e3c2df2ef4d-buildcachedir\") pod \"sg-core-1-build\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " pod="service-telemetry/sg-core-1-build" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.258164 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/845832d9-625c-452e-b900-4e3c2df2ef4d-container-storage-root\") pod \"sg-core-1-build\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " pod="service-telemetry/sg-core-1-build" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.258183 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/845832d9-625c-452e-b900-4e3c2df2ef4d-node-pullsecrets\") pod \"sg-core-1-build\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " pod="service-telemetry/sg-core-1-build" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.258207 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/845832d9-625c-452e-b900-4e3c2df2ef4d-build-proxy-ca-bundles\") pod \"sg-core-1-build\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " pod="service-telemetry/sg-core-1-build" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.258244 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/845832d9-625c-452e-b900-4e3c2df2ef4d-builder-dockercfg-8h4xs-push\") pod \"sg-core-1-build\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " pod="service-telemetry/sg-core-1-build" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.258266 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/845832d9-625c-452e-b900-4e3c2df2ef4d-build-ca-bundles\") pod \"sg-core-1-build\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " pod="service-telemetry/sg-core-1-build" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.258280 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/845832d9-625c-452e-b900-4e3c2df2ef4d-container-storage-run\") pod \"sg-core-1-build\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " pod="service-telemetry/sg-core-1-build" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.258298 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/845832d9-625c-452e-b900-4e3c2df2ef4d-buildworkdir\") pod \"sg-core-1-build\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " pod="service-telemetry/sg-core-1-build" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.258311 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/845832d9-625c-452e-b900-4e3c2df2ef4d-builder-dockercfg-8h4xs-pull\") pod \"sg-core-1-build\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " pod="service-telemetry/sg-core-1-build" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.258791 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/845832d9-625c-452e-b900-4e3c2df2ef4d-build-system-configs\") pod \"sg-core-1-build\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " pod="service-telemetry/sg-core-1-build" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.258822 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/845832d9-625c-452e-b900-4e3c2df2ef4d-node-pullsecrets\") pod \"sg-core-1-build\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " pod="service-telemetry/sg-core-1-build" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.258978 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/845832d9-625c-452e-b900-4e3c2df2ef4d-container-storage-root\") pod \"sg-core-1-build\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " pod="service-telemetry/sg-core-1-build" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.259043 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/845832d9-625c-452e-b900-4e3c2df2ef4d-container-storage-run\") pod \"sg-core-1-build\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " pod="service-telemetry/sg-core-1-build" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.259044 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/845832d9-625c-452e-b900-4e3c2df2ef4d-buildcachedir\") pod \"sg-core-1-build\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " pod="service-telemetry/sg-core-1-build" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.259224 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/845832d9-625c-452e-b900-4e3c2df2ef4d-buildworkdir\") pod \"sg-core-1-build\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " pod="service-telemetry/sg-core-1-build" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.259703 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/845832d9-625c-452e-b900-4e3c2df2ef4d-build-proxy-ca-bundles\") pod \"sg-core-1-build\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " pod="service-telemetry/sg-core-1-build" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.259964 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/845832d9-625c-452e-b900-4e3c2df2ef4d-build-blob-cache\") pod \"sg-core-1-build\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " pod="service-telemetry/sg-core-1-build" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.259972 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/845832d9-625c-452e-b900-4e3c2df2ef4d-build-ca-bundles\") pod \"sg-core-1-build\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " pod="service-telemetry/sg-core-1-build" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.265567 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/845832d9-625c-452e-b900-4e3c2df2ef4d-builder-dockercfg-8h4xs-push\") pod \"sg-core-1-build\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " pod="service-telemetry/sg-core-1-build" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.270192 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/845832d9-625c-452e-b900-4e3c2df2ef4d-builder-dockercfg-8h4xs-pull\") pod \"sg-core-1-build\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " pod="service-telemetry/sg-core-1-build" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.282477 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7glkd\" (UniqueName: \"kubernetes.io/projected/845832d9-625c-452e-b900-4e3c2df2ef4d-kube-api-access-7glkd\") pod \"sg-core-1-build\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " pod="service-telemetry/sg-core-1-build" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.330047 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-1-build" Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.739282 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-core-1-build"] Jan 26 17:17:09 crc kubenswrapper[4856]: I0126 17:17:09.782645 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-1-build" event={"ID":"845832d9-625c-452e-b900-4e3c2df2ef4d","Type":"ContainerStarted","Data":"37f3f15c961670bf793cb2d570bf81f5b90a0fceb924638d55df34f9da6e88b4"} Jan 26 17:17:10 crc kubenswrapper[4856]: I0126 17:17:10.790403 4856 generic.go:334] "Generic (PLEG): container finished" podID="845832d9-625c-452e-b900-4e3c2df2ef4d" containerID="86acff0987e388de3d9ce01fcd9e358d21cc1ecfc613874a4c26f37c1a31ec0b" exitCode=0 Jan 26 17:17:10 crc kubenswrapper[4856]: I0126 17:17:10.790557 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-1-build" event={"ID":"845832d9-625c-452e-b900-4e3c2df2ef4d","Type":"ContainerDied","Data":"86acff0987e388de3d9ce01fcd9e358d21cc1ecfc613874a4c26f37c1a31ec0b"} Jan 26 17:17:11 crc kubenswrapper[4856]: I0126 17:17:11.799899 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-1-build" event={"ID":"845832d9-625c-452e-b900-4e3c2df2ef4d","Type":"ContainerStarted","Data":"00702564f2498cf2bf404ba97ad4e9284443fddbcee3e0fe6e8490fc7acce163"} Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.192803 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/sg-core-1-build" podStartSLOduration=11.192765158 podStartE2EDuration="11.192765158s" podCreationTimestamp="2026-01-26 17:17:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:17:11.828766028 +0000 UTC m=+1127.782020009" watchObservedRunningTime="2026-01-26 17:17:19.192765158 +0000 UTC m=+1135.146019139" Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.197372 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/sg-core-1-build"] Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.197701 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/sg-core-1-build" podUID="845832d9-625c-452e-b900-4e3c2df2ef4d" containerName="docker-build" containerID="cri-o://00702564f2498cf2bf404ba97ad4e9284443fddbcee3e0fe6e8490fc7acce163" gracePeriod=30 Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.617418 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-core-1-build_845832d9-625c-452e-b900-4e3c2df2ef4d/docker-build/0.log" Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.618476 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-1-build" Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.875654 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/845832d9-625c-452e-b900-4e3c2df2ef4d-buildcachedir\") pod \"845832d9-625c-452e-b900-4e3c2df2ef4d\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.875789 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/845832d9-625c-452e-b900-4e3c2df2ef4d-container-storage-root\") pod \"845832d9-625c-452e-b900-4e3c2df2ef4d\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.875840 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/845832d9-625c-452e-b900-4e3c2df2ef4d-build-blob-cache\") pod \"845832d9-625c-452e-b900-4e3c2df2ef4d\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.875925 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/845832d9-625c-452e-b900-4e3c2df2ef4d-container-storage-run\") pod \"845832d9-625c-452e-b900-4e3c2df2ef4d\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.875969 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/845832d9-625c-452e-b900-4e3c2df2ef4d-build-ca-bundles\") pod \"845832d9-625c-452e-b900-4e3c2df2ef4d\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.876014 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/845832d9-625c-452e-b900-4e3c2df2ef4d-builder-dockercfg-8h4xs-push\") pod \"845832d9-625c-452e-b900-4e3c2df2ef4d\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.876137 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/845832d9-625c-452e-b900-4e3c2df2ef4d-builder-dockercfg-8h4xs-pull\") pod \"845832d9-625c-452e-b900-4e3c2df2ef4d\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.876169 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/845832d9-625c-452e-b900-4e3c2df2ef4d-build-system-configs\") pod \"845832d9-625c-452e-b900-4e3c2df2ef4d\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.876198 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7glkd\" (UniqueName: \"kubernetes.io/projected/845832d9-625c-452e-b900-4e3c2df2ef4d-kube-api-access-7glkd\") pod \"845832d9-625c-452e-b900-4e3c2df2ef4d\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.876267 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/845832d9-625c-452e-b900-4e3c2df2ef4d-node-pullsecrets\") pod \"845832d9-625c-452e-b900-4e3c2df2ef4d\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.876352 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/845832d9-625c-452e-b900-4e3c2df2ef4d-build-proxy-ca-bundles\") pod \"845832d9-625c-452e-b900-4e3c2df2ef4d\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.876398 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/845832d9-625c-452e-b900-4e3c2df2ef4d-buildworkdir\") pod \"845832d9-625c-452e-b900-4e3c2df2ef4d\" (UID: \"845832d9-625c-452e-b900-4e3c2df2ef4d\") " Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.878425 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/845832d9-625c-452e-b900-4e3c2df2ef4d-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "845832d9-625c-452e-b900-4e3c2df2ef4d" (UID: "845832d9-625c-452e-b900-4e3c2df2ef4d"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.878607 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/845832d9-625c-452e-b900-4e3c2df2ef4d-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "845832d9-625c-452e-b900-4e3c2df2ef4d" (UID: "845832d9-625c-452e-b900-4e3c2df2ef4d"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.879275 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/845832d9-625c-452e-b900-4e3c2df2ef4d-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "845832d9-625c-452e-b900-4e3c2df2ef4d" (UID: "845832d9-625c-452e-b900-4e3c2df2ef4d"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.881774 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/845832d9-625c-452e-b900-4e3c2df2ef4d-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "845832d9-625c-452e-b900-4e3c2df2ef4d" (UID: "845832d9-625c-452e-b900-4e3c2df2ef4d"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.882256 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/845832d9-625c-452e-b900-4e3c2df2ef4d-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "845832d9-625c-452e-b900-4e3c2df2ef4d" (UID: "845832d9-625c-452e-b900-4e3c2df2ef4d"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.882617 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/845832d9-625c-452e-b900-4e3c2df2ef4d-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "845832d9-625c-452e-b900-4e3c2df2ef4d" (UID: "845832d9-625c-452e-b900-4e3c2df2ef4d"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.885752 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/845832d9-625c-452e-b900-4e3c2df2ef4d-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "845832d9-625c-452e-b900-4e3c2df2ef4d" (UID: "845832d9-625c-452e-b900-4e3c2df2ef4d"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.885780 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/845832d9-625c-452e-b900-4e3c2df2ef4d-builder-dockercfg-8h4xs-push" (OuterVolumeSpecName: "builder-dockercfg-8h4xs-push") pod "845832d9-625c-452e-b900-4e3c2df2ef4d" (UID: "845832d9-625c-452e-b900-4e3c2df2ef4d"). InnerVolumeSpecName "builder-dockercfg-8h4xs-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.889883 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-core-1-build_845832d9-625c-452e-b900-4e3c2df2ef4d/docker-build/0.log" Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.893397 4856 generic.go:334] "Generic (PLEG): container finished" podID="845832d9-625c-452e-b900-4e3c2df2ef4d" containerID="00702564f2498cf2bf404ba97ad4e9284443fddbcee3e0fe6e8490fc7acce163" exitCode=1 Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.893463 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-1-build" event={"ID":"845832d9-625c-452e-b900-4e3c2df2ef4d","Type":"ContainerDied","Data":"00702564f2498cf2bf404ba97ad4e9284443fddbcee3e0fe6e8490fc7acce163"} Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.893508 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-1-build" event={"ID":"845832d9-625c-452e-b900-4e3c2df2ef4d","Type":"ContainerDied","Data":"37f3f15c961670bf793cb2d570bf81f5b90a0fceb924638d55df34f9da6e88b4"} Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.893587 4856 scope.go:117] "RemoveContainer" containerID="00702564f2498cf2bf404ba97ad4e9284443fddbcee3e0fe6e8490fc7acce163" Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.893800 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-1-build" Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.895785 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/845832d9-625c-452e-b900-4e3c2df2ef4d-builder-dockercfg-8h4xs-pull" (OuterVolumeSpecName: "builder-dockercfg-8h4xs-pull") pod "845832d9-625c-452e-b900-4e3c2df2ef4d" (UID: "845832d9-625c-452e-b900-4e3c2df2ef4d"). InnerVolumeSpecName "builder-dockercfg-8h4xs-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.910072 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/845832d9-625c-452e-b900-4e3c2df2ef4d-kube-api-access-7glkd" (OuterVolumeSpecName: "kube-api-access-7glkd") pod "845832d9-625c-452e-b900-4e3c2df2ef4d" (UID: "845832d9-625c-452e-b900-4e3c2df2ef4d"). InnerVolumeSpecName "kube-api-access-7glkd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.970361 4856 scope.go:117] "RemoveContainer" containerID="86acff0987e388de3d9ce01fcd9e358d21cc1ecfc613874a4c26f37c1a31ec0b" Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.979219 4856 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/845832d9-625c-452e-b900-4e3c2df2ef4d-buildcachedir\") on node \"crc\" DevicePath \"\"" Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.979280 4856 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/845832d9-625c-452e-b900-4e3c2df2ef4d-container-storage-run\") on node \"crc\" DevicePath \"\"" Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.979367 4856 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/845832d9-625c-452e-b900-4e3c2df2ef4d-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.979382 4856 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/845832d9-625c-452e-b900-4e3c2df2ef4d-builder-dockercfg-8h4xs-push\") on node \"crc\" DevicePath \"\"" Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.979418 4856 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/845832d9-625c-452e-b900-4e3c2df2ef4d-builder-dockercfg-8h4xs-pull\") on node \"crc\" DevicePath \"\"" Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.979432 4856 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/845832d9-625c-452e-b900-4e3c2df2ef4d-build-system-configs\") on node \"crc\" DevicePath \"\"" Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.979440 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7glkd\" (UniqueName: \"kubernetes.io/projected/845832d9-625c-452e-b900-4e3c2df2ef4d-kube-api-access-7glkd\") on node \"crc\" DevicePath \"\"" Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.979453 4856 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/845832d9-625c-452e-b900-4e3c2df2ef4d-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.979461 4856 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/845832d9-625c-452e-b900-4e3c2df2ef4d-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 26 17:17:19 crc kubenswrapper[4856]: I0126 17:17:19.979471 4856 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/845832d9-625c-452e-b900-4e3c2df2ef4d-buildworkdir\") on node \"crc\" DevicePath \"\"" Jan 26 17:17:20 crc kubenswrapper[4856]: I0126 17:17:20.000615 4856 scope.go:117] "RemoveContainer" containerID="00702564f2498cf2bf404ba97ad4e9284443fddbcee3e0fe6e8490fc7acce163" Jan 26 17:17:20 crc kubenswrapper[4856]: E0126 17:17:20.001342 4856 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00702564f2498cf2bf404ba97ad4e9284443fddbcee3e0fe6e8490fc7acce163\": container with ID starting with 00702564f2498cf2bf404ba97ad4e9284443fddbcee3e0fe6e8490fc7acce163 not found: ID does not exist" containerID="00702564f2498cf2bf404ba97ad4e9284443fddbcee3e0fe6e8490fc7acce163" Jan 26 17:17:20 crc kubenswrapper[4856]: I0126 17:17:20.001455 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00702564f2498cf2bf404ba97ad4e9284443fddbcee3e0fe6e8490fc7acce163"} err="failed to get container status \"00702564f2498cf2bf404ba97ad4e9284443fddbcee3e0fe6e8490fc7acce163\": rpc error: code = NotFound desc = could not find container \"00702564f2498cf2bf404ba97ad4e9284443fddbcee3e0fe6e8490fc7acce163\": container with ID starting with 00702564f2498cf2bf404ba97ad4e9284443fddbcee3e0fe6e8490fc7acce163 not found: ID does not exist" Jan 26 17:17:20 crc kubenswrapper[4856]: I0126 17:17:20.001514 4856 scope.go:117] "RemoveContainer" containerID="86acff0987e388de3d9ce01fcd9e358d21cc1ecfc613874a4c26f37c1a31ec0b" Jan 26 17:17:20 crc kubenswrapper[4856]: E0126 17:17:20.002065 4856 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86acff0987e388de3d9ce01fcd9e358d21cc1ecfc613874a4c26f37c1a31ec0b\": container with ID starting with 86acff0987e388de3d9ce01fcd9e358d21cc1ecfc613874a4c26f37c1a31ec0b not found: ID does not exist" containerID="86acff0987e388de3d9ce01fcd9e358d21cc1ecfc613874a4c26f37c1a31ec0b" Jan 26 17:17:20 crc kubenswrapper[4856]: I0126 17:17:20.002104 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86acff0987e388de3d9ce01fcd9e358d21cc1ecfc613874a4c26f37c1a31ec0b"} err="failed to get container status \"86acff0987e388de3d9ce01fcd9e358d21cc1ecfc613874a4c26f37c1a31ec0b\": rpc error: code = NotFound desc = could not find container \"86acff0987e388de3d9ce01fcd9e358d21cc1ecfc613874a4c26f37c1a31ec0b\": container with ID starting with 86acff0987e388de3d9ce01fcd9e358d21cc1ecfc613874a4c26f37c1a31ec0b not found: ID does not exist" Jan 26 17:17:20 crc kubenswrapper[4856]: I0126 17:17:20.005682 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/845832d9-625c-452e-b900-4e3c2df2ef4d-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "845832d9-625c-452e-b900-4e3c2df2ef4d" (UID: "845832d9-625c-452e-b900-4e3c2df2ef4d"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:17:20 crc kubenswrapper[4856]: I0126 17:17:20.055499 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/845832d9-625c-452e-b900-4e3c2df2ef4d-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "845832d9-625c-452e-b900-4e3c2df2ef4d" (UID: "845832d9-625c-452e-b900-4e3c2df2ef4d"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:17:20 crc kubenswrapper[4856]: I0126 17:17:20.080629 4856 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/845832d9-625c-452e-b900-4e3c2df2ef4d-container-storage-root\") on node \"crc\" DevicePath \"\"" Jan 26 17:17:20 crc kubenswrapper[4856]: I0126 17:17:20.080707 4856 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/845832d9-625c-452e-b900-4e3c2df2ef4d-build-blob-cache\") on node \"crc\" DevicePath \"\"" Jan 26 17:17:20 crc kubenswrapper[4856]: I0126 17:17:20.228728 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/sg-core-1-build"] Jan 26 17:17:20 crc kubenswrapper[4856]: I0126 17:17:20.235980 4856 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/sg-core-1-build"] Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.074009 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/sg-core-2-build"] Jan 26 17:17:21 crc kubenswrapper[4856]: E0126 17:17:21.074631 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="845832d9-625c-452e-b900-4e3c2df2ef4d" containerName="docker-build" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.074652 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="845832d9-625c-452e-b900-4e3c2df2ef4d" containerName="docker-build" Jan 26 17:17:21 crc kubenswrapper[4856]: E0126 17:17:21.074667 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="845832d9-625c-452e-b900-4e3c2df2ef4d" containerName="manage-dockerfile" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.074674 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="845832d9-625c-452e-b900-4e3c2df2ef4d" containerName="manage-dockerfile" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.074855 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="845832d9-625c-452e-b900-4e3c2df2ef4d" containerName="docker-build" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.077587 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-2-build" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.080998 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-core-2-sys-config" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.081050 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-core-2-ca" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.081482 4856 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-8h4xs" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.081620 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-core-2-global-ca" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.094374 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-core-2-build"] Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.115739 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vg4d\" (UniqueName: \"kubernetes.io/projected/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-kube-api-access-9vg4d\") pod \"sg-core-2-build\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " pod="service-telemetry/sg-core-2-build" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.116252 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-build-system-configs\") pod \"sg-core-2-build\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " pod="service-telemetry/sg-core-2-build" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.116409 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-container-storage-root\") pod \"sg-core-2-build\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " pod="service-telemetry/sg-core-2-build" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.116750 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-build-blob-cache\") pod \"sg-core-2-build\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " pod="service-telemetry/sg-core-2-build" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.116844 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-build-proxy-ca-bundles\") pod \"sg-core-2-build\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " pod="service-telemetry/sg-core-2-build" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.117045 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-buildcachedir\") pod \"sg-core-2-build\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " pod="service-telemetry/sg-core-2-build" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.117143 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-builder-dockercfg-8h4xs-push\") pod \"sg-core-2-build\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " pod="service-telemetry/sg-core-2-build" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.117229 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-buildworkdir\") pod \"sg-core-2-build\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " pod="service-telemetry/sg-core-2-build" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.117369 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-container-storage-run\") pod \"sg-core-2-build\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " pod="service-telemetry/sg-core-2-build" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.117512 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-build-ca-bundles\") pod \"sg-core-2-build\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " pod="service-telemetry/sg-core-2-build" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.117616 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-builder-dockercfg-8h4xs-pull\") pod \"sg-core-2-build\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " pod="service-telemetry/sg-core-2-build" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.117723 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-node-pullsecrets\") pod \"sg-core-2-build\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " pod="service-telemetry/sg-core-2-build" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.219654 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vg4d\" (UniqueName: \"kubernetes.io/projected/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-kube-api-access-9vg4d\") pod \"sg-core-2-build\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " pod="service-telemetry/sg-core-2-build" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.220569 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-build-system-configs\") pod \"sg-core-2-build\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " pod="service-telemetry/sg-core-2-build" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.220728 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-container-storage-root\") pod \"sg-core-2-build\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " pod="service-telemetry/sg-core-2-build" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.220824 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-build-blob-cache\") pod \"sg-core-2-build\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " pod="service-telemetry/sg-core-2-build" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.220876 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-build-proxy-ca-bundles\") pod \"sg-core-2-build\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " pod="service-telemetry/sg-core-2-build" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.220936 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-buildcachedir\") pod \"sg-core-2-build\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " pod="service-telemetry/sg-core-2-build" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.221015 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-builder-dockercfg-8h4xs-push\") pod \"sg-core-2-build\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " pod="service-telemetry/sg-core-2-build" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.221081 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-buildworkdir\") pod \"sg-core-2-build\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " pod="service-telemetry/sg-core-2-build" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.221104 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-container-storage-run\") pod \"sg-core-2-build\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " pod="service-telemetry/sg-core-2-build" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.221125 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-build-ca-bundles\") pod \"sg-core-2-build\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " pod="service-telemetry/sg-core-2-build" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.221141 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-build-blob-cache\") pod \"sg-core-2-build\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " pod="service-telemetry/sg-core-2-build" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.221153 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-builder-dockercfg-8h4xs-pull\") pod \"sg-core-2-build\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " pod="service-telemetry/sg-core-2-build" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.221188 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-container-storage-root\") pod \"sg-core-2-build\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " pod="service-telemetry/sg-core-2-build" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.221235 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-node-pullsecrets\") pod \"sg-core-2-build\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " pod="service-telemetry/sg-core-2-build" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.221361 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-node-pullsecrets\") pod \"sg-core-2-build\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " pod="service-telemetry/sg-core-2-build" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.221453 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-build-system-configs\") pod \"sg-core-2-build\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " pod="service-telemetry/sg-core-2-build" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.221594 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-buildcachedir\") pod \"sg-core-2-build\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " pod="service-telemetry/sg-core-2-build" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.221688 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-buildworkdir\") pod \"sg-core-2-build\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " pod="service-telemetry/sg-core-2-build" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.221861 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-container-storage-run\") pod \"sg-core-2-build\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " pod="service-telemetry/sg-core-2-build" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.221905 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-build-proxy-ca-bundles\") pod \"sg-core-2-build\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " pod="service-telemetry/sg-core-2-build" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.222358 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-build-ca-bundles\") pod \"sg-core-2-build\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " pod="service-telemetry/sg-core-2-build" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.225982 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-builder-dockercfg-8h4xs-push\") pod \"sg-core-2-build\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " pod="service-telemetry/sg-core-2-build" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.226049 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-builder-dockercfg-8h4xs-pull\") pod \"sg-core-2-build\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " pod="service-telemetry/sg-core-2-build" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.240757 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vg4d\" (UniqueName: \"kubernetes.io/projected/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-kube-api-access-9vg4d\") pod \"sg-core-2-build\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " pod="service-telemetry/sg-core-2-build" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.404824 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="845832d9-625c-452e-b900-4e3c2df2ef4d" path="/var/lib/kubelet/pods/845832d9-625c-452e-b900-4e3c2df2ef4d/volumes" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.442953 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-2-build" Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.667444 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-core-2-build"] Jan 26 17:17:21 crc kubenswrapper[4856]: I0126 17:17:21.907673 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906","Type":"ContainerStarted","Data":"581a54ce580bf89c79c4ca5c090533d91b572d9478534f85f14b11f0a695bf7c"} Jan 26 17:17:22 crc kubenswrapper[4856]: I0126 17:17:22.918833 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906","Type":"ContainerStarted","Data":"c3d2079125adda4fff730760ad4a3c9ff84f0c4bb91f73b927cd7c20ec1365fc"} Jan 26 17:17:24 crc kubenswrapper[4856]: I0126 17:17:24.105442 4856 generic.go:334] "Generic (PLEG): container finished" podID="230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906" containerID="c3d2079125adda4fff730760ad4a3c9ff84f0c4bb91f73b927cd7c20ec1365fc" exitCode=0 Jan 26 17:17:24 crc kubenswrapper[4856]: I0126 17:17:24.106448 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906","Type":"ContainerDied","Data":"c3d2079125adda4fff730760ad4a3c9ff84f0c4bb91f73b927cd7c20ec1365fc"} Jan 26 17:17:25 crc kubenswrapper[4856]: I0126 17:17:25.115448 4856 generic.go:334] "Generic (PLEG): container finished" podID="230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906" containerID="1e916af40f386b7c7eeeca4a494f1a4ea6edb3f8a20e51f7232fa3f4df21515f" exitCode=0 Jan 26 17:17:25 crc kubenswrapper[4856]: I0126 17:17:25.115505 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906","Type":"ContainerDied","Data":"1e916af40f386b7c7eeeca4a494f1a4ea6edb3f8a20e51f7232fa3f4df21515f"} Jan 26 17:17:25 crc kubenswrapper[4856]: I0126 17:17:25.165385 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-core-2-build_230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906/manage-dockerfile/0.log" Jan 26 17:17:26 crc kubenswrapper[4856]: I0126 17:17:26.126808 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906","Type":"ContainerStarted","Data":"0c8b0cbcc314d24ca3abe306e12646e8cc81d7ef948190b4214ea9f83847d71e"} Jan 26 17:17:26 crc kubenswrapper[4856]: I0126 17:17:26.165656 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/sg-core-2-build" podStartSLOduration=5.165640123 podStartE2EDuration="5.165640123s" podCreationTimestamp="2026-01-26 17:17:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:17:26.165225321 +0000 UTC m=+1142.118479332" watchObservedRunningTime="2026-01-26 17:17:26.165640123 +0000 UTC m=+1142.118894104" Jan 26 17:18:56 crc kubenswrapper[4856]: I0126 17:18:56.939229 4856 patch_prober.go:28] interesting pod/machine-config-daemon-xm9cq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:18:56 crc kubenswrapper[4856]: I0126 17:18:56.940159 4856 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:19:26 crc kubenswrapper[4856]: I0126 17:19:26.938865 4856 patch_prober.go:28] interesting pod/machine-config-daemon-xm9cq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:19:26 crc kubenswrapper[4856]: I0126 17:19:26.939459 4856 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:19:56 crc kubenswrapper[4856]: I0126 17:19:56.938637 4856 patch_prober.go:28] interesting pod/machine-config-daemon-xm9cq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:19:56 crc kubenswrapper[4856]: I0126 17:19:56.939238 4856 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:19:56 crc kubenswrapper[4856]: I0126 17:19:56.939296 4856 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" Jan 26 17:19:56 crc kubenswrapper[4856]: I0126 17:19:56.940145 4856 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5846ab4d870be5fcbab796c3e27690d2c13d129480d6fcd21b3b0d1c535f0cff"} pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 26 17:19:56 crc kubenswrapper[4856]: I0126 17:19:56.940232 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" containerName="machine-config-daemon" containerID="cri-o://5846ab4d870be5fcbab796c3e27690d2c13d129480d6fcd21b3b0d1c535f0cff" gracePeriod=600 Jan 26 17:19:57 crc kubenswrapper[4856]: I0126 17:19:57.851713 4856 generic.go:334] "Generic (PLEG): container finished" podID="63c75ede-5170-4db0-811b-5217ef8d72b3" containerID="5846ab4d870be5fcbab796c3e27690d2c13d129480d6fcd21b3b0d1c535f0cff" exitCode=0 Jan 26 17:19:57 crc kubenswrapper[4856]: I0126 17:19:57.851747 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" event={"ID":"63c75ede-5170-4db0-811b-5217ef8d72b3","Type":"ContainerDied","Data":"5846ab4d870be5fcbab796c3e27690d2c13d129480d6fcd21b3b0d1c535f0cff"} Jan 26 17:19:57 crc kubenswrapper[4856]: I0126 17:19:57.852565 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" event={"ID":"63c75ede-5170-4db0-811b-5217ef8d72b3","Type":"ContainerStarted","Data":"cda3cdbac0b1e3c460ee9a5617b9c5fd59d4db5c67a69b81c9224934be12563c"} Jan 26 17:19:57 crc kubenswrapper[4856]: I0126 17:19:57.852640 4856 scope.go:117] "RemoveContainer" containerID="fdaad4602089daad40b0395fbc761e615a8ba2a94c8f5b977142a787034cddb7" Jan 26 17:20:46 crc kubenswrapper[4856]: E0126 17:20:46.997977 4856 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.604s" Jan 26 17:21:19 crc kubenswrapper[4856]: I0126 17:21:19.253610 4856 generic.go:334] "Generic (PLEG): container finished" podID="230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906" containerID="0c8b0cbcc314d24ca3abe306e12646e8cc81d7ef948190b4214ea9f83847d71e" exitCode=0 Jan 26 17:21:19 crc kubenswrapper[4856]: I0126 17:21:19.253748 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906","Type":"ContainerDied","Data":"0c8b0cbcc314d24ca3abe306e12646e8cc81d7ef948190b4214ea9f83847d71e"} Jan 26 17:21:20 crc kubenswrapper[4856]: I0126 17:21:20.564739 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-2-build" Jan 26 17:21:20 crc kubenswrapper[4856]: I0126 17:21:20.686923 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-build-ca-bundles\") pod \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " Jan 26 17:21:20 crc kubenswrapper[4856]: I0126 17:21:20.686996 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-buildworkdir\") pod \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " Jan 26 17:21:20 crc kubenswrapper[4856]: I0126 17:21:20.687024 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-container-storage-root\") pod \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " Jan 26 17:21:20 crc kubenswrapper[4856]: I0126 17:21:20.687083 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-builder-dockercfg-8h4xs-pull\") pod \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " Jan 26 17:21:20 crc kubenswrapper[4856]: I0126 17:21:20.687120 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-builder-dockercfg-8h4xs-push\") pod \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " Jan 26 17:21:20 crc kubenswrapper[4856]: I0126 17:21:20.687146 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-build-system-configs\") pod \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " Jan 26 17:21:20 crc kubenswrapper[4856]: I0126 17:21:20.687169 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-build-blob-cache\") pod \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " Jan 26 17:21:20 crc kubenswrapper[4856]: I0126 17:21:20.687205 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-container-storage-run\") pod \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " Jan 26 17:21:20 crc kubenswrapper[4856]: I0126 17:21:20.687242 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-node-pullsecrets\") pod \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " Jan 26 17:21:20 crc kubenswrapper[4856]: I0126 17:21:20.687264 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-build-proxy-ca-bundles\") pod \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " Jan 26 17:21:20 crc kubenswrapper[4856]: I0126 17:21:20.687288 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9vg4d\" (UniqueName: \"kubernetes.io/projected/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-kube-api-access-9vg4d\") pod \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " Jan 26 17:21:20 crc kubenswrapper[4856]: I0126 17:21:20.687333 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-buildcachedir\") pod \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\" (UID: \"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906\") " Jan 26 17:21:20 crc kubenswrapper[4856]: I0126 17:21:20.687600 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906" (UID: "230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:21:20 crc kubenswrapper[4856]: I0126 17:21:20.687640 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906" (UID: "230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:21:20 crc kubenswrapper[4856]: I0126 17:21:20.689232 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906" (UID: "230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:21:20 crc kubenswrapper[4856]: I0126 17:21:20.689244 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906" (UID: "230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:21:20 crc kubenswrapper[4856]: I0126 17:21:20.690328 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906" (UID: "230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:21:20 crc kubenswrapper[4856]: I0126 17:21:20.690764 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906" (UID: "230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:21:20 crc kubenswrapper[4856]: I0126 17:21:20.693408 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-builder-dockercfg-8h4xs-pull" (OuterVolumeSpecName: "builder-dockercfg-8h4xs-pull") pod "230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906" (UID: "230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906"). InnerVolumeSpecName "builder-dockercfg-8h4xs-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:21:20 crc kubenswrapper[4856]: I0126 17:21:20.693427 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-builder-dockercfg-8h4xs-push" (OuterVolumeSpecName: "builder-dockercfg-8h4xs-push") pod "230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906" (UID: "230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906"). InnerVolumeSpecName "builder-dockercfg-8h4xs-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:21:20 crc kubenswrapper[4856]: I0126 17:21:20.703187 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-kube-api-access-9vg4d" (OuterVolumeSpecName: "kube-api-access-9vg4d") pod "230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906" (UID: "230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906"). InnerVolumeSpecName "kube-api-access-9vg4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:21:20 crc kubenswrapper[4856]: I0126 17:21:20.713256 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906" (UID: "230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:21:20 crc kubenswrapper[4856]: I0126 17:21:20.788733 4856 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-builder-dockercfg-8h4xs-push\") on node \"crc\" DevicePath \"\"" Jan 26 17:21:20 crc kubenswrapper[4856]: I0126 17:21:20.788775 4856 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-build-system-configs\") on node \"crc\" DevicePath \"\"" Jan 26 17:21:20 crc kubenswrapper[4856]: I0126 17:21:20.788789 4856 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-container-storage-run\") on node \"crc\" DevicePath \"\"" Jan 26 17:21:20 crc kubenswrapper[4856]: I0126 17:21:20.788803 4856 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Jan 26 17:21:20 crc kubenswrapper[4856]: I0126 17:21:20.788816 4856 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 26 17:21:20 crc kubenswrapper[4856]: I0126 17:21:20.788827 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9vg4d\" (UniqueName: \"kubernetes.io/projected/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-kube-api-access-9vg4d\") on node \"crc\" DevicePath \"\"" Jan 26 17:21:20 crc kubenswrapper[4856]: I0126 17:21:20.788838 4856 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-buildcachedir\") on node \"crc\" DevicePath \"\"" Jan 26 17:21:20 crc kubenswrapper[4856]: I0126 17:21:20.788850 4856 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 26 17:21:20 crc kubenswrapper[4856]: I0126 17:21:20.788862 4856 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-buildworkdir\") on node \"crc\" DevicePath \"\"" Jan 26 17:21:20 crc kubenswrapper[4856]: I0126 17:21:20.788875 4856 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-builder-dockercfg-8h4xs-pull\") on node \"crc\" DevicePath \"\"" Jan 26 17:21:21 crc kubenswrapper[4856]: I0126 17:21:21.008433 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906" (UID: "230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:21:21 crc kubenswrapper[4856]: I0126 17:21:21.093289 4856 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-build-blob-cache\") on node \"crc\" DevicePath \"\"" Jan 26 17:21:21 crc kubenswrapper[4856]: I0126 17:21:21.269509 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906","Type":"ContainerDied","Data":"581a54ce580bf89c79c4ca5c090533d91b572d9478534f85f14b11f0a695bf7c"} Jan 26 17:21:21 crc kubenswrapper[4856]: I0126 17:21:21.269589 4856 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="581a54ce580bf89c79c4ca5c090533d91b572d9478534f85f14b11f0a695bf7c" Jan 26 17:21:21 crc kubenswrapper[4856]: I0126 17:21:21.269915 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-2-build" Jan 26 17:21:23 crc kubenswrapper[4856]: I0126 17:21:23.156301 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906" (UID: "230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:21:23 crc kubenswrapper[4856]: I0126 17:21:23.225532 4856 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906-container-storage-root\") on node \"crc\" DevicePath \"\"" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.051504 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/sg-bridge-1-build"] Jan 26 17:21:26 crc kubenswrapper[4856]: E0126 17:21:26.052239 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906" containerName="docker-build" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.052264 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906" containerName="docker-build" Jan 26 17:21:26 crc kubenswrapper[4856]: E0126 17:21:26.052283 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906" containerName="manage-dockerfile" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.052292 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906" containerName="manage-dockerfile" Jan 26 17:21:26 crc kubenswrapper[4856]: E0126 17:21:26.052308 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906" containerName="git-clone" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.052317 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906" containerName="git-clone" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.052547 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="230fa7ab-5d3d-40d2-bcf1-6ee5a68a3906" containerName="docker-build" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.053437 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-1-build" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.055567 4856 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-8h4xs" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.055849 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-bridge-1-ca" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.059043 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-bridge-1-sys-config" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.059671 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-bridge-1-global-ca" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.071486 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-bridge-1-build"] Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.169968 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d55d4558-c971-4a40-b1cb-41389cbf11c3-node-pullsecrets\") pod \"sg-bridge-1-build\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " pod="service-telemetry/sg-bridge-1-build" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.170035 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpgwg\" (UniqueName: \"kubernetes.io/projected/d55d4558-c971-4a40-b1cb-41389cbf11c3-kube-api-access-jpgwg\") pod \"sg-bridge-1-build\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " pod="service-telemetry/sg-bridge-1-build" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.170071 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/d55d4558-c971-4a40-b1cb-41389cbf11c3-buildcachedir\") pod \"sg-bridge-1-build\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " pod="service-telemetry/sg-bridge-1-build" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.170109 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/d55d4558-c971-4a40-b1cb-41389cbf11c3-buildworkdir\") pod \"sg-bridge-1-build\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " pod="service-telemetry/sg-bridge-1-build" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.170229 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/d55d4558-c971-4a40-b1cb-41389cbf11c3-build-system-configs\") pod \"sg-bridge-1-build\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " pod="service-telemetry/sg-bridge-1-build" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.170313 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/d55d4558-c971-4a40-b1cb-41389cbf11c3-container-storage-run\") pod \"sg-bridge-1-build\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " pod="service-telemetry/sg-bridge-1-build" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.170344 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/d55d4558-c971-4a40-b1cb-41389cbf11c3-build-blob-cache\") pod \"sg-bridge-1-build\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " pod="service-telemetry/sg-bridge-1-build" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.170382 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d55d4558-c971-4a40-b1cb-41389cbf11c3-build-proxy-ca-bundles\") pod \"sg-bridge-1-build\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " pod="service-telemetry/sg-bridge-1-build" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.170483 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d55d4558-c971-4a40-b1cb-41389cbf11c3-build-ca-bundles\") pod \"sg-bridge-1-build\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " pod="service-telemetry/sg-bridge-1-build" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.170550 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/d55d4558-c971-4a40-b1cb-41389cbf11c3-builder-dockercfg-8h4xs-push\") pod \"sg-bridge-1-build\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " pod="service-telemetry/sg-bridge-1-build" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.170587 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/d55d4558-c971-4a40-b1cb-41389cbf11c3-builder-dockercfg-8h4xs-pull\") pod \"sg-bridge-1-build\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " pod="service-telemetry/sg-bridge-1-build" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.170607 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/d55d4558-c971-4a40-b1cb-41389cbf11c3-container-storage-root\") pod \"sg-bridge-1-build\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " pod="service-telemetry/sg-bridge-1-build" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.271897 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpgwg\" (UniqueName: \"kubernetes.io/projected/d55d4558-c971-4a40-b1cb-41389cbf11c3-kube-api-access-jpgwg\") pod \"sg-bridge-1-build\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " pod="service-telemetry/sg-bridge-1-build" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.271953 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/d55d4558-c971-4a40-b1cb-41389cbf11c3-buildcachedir\") pod \"sg-bridge-1-build\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " pod="service-telemetry/sg-bridge-1-build" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.271998 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/d55d4558-c971-4a40-b1cb-41389cbf11c3-buildworkdir\") pod \"sg-bridge-1-build\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " pod="service-telemetry/sg-bridge-1-build" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.272022 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/d55d4558-c971-4a40-b1cb-41389cbf11c3-build-system-configs\") pod \"sg-bridge-1-build\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " pod="service-telemetry/sg-bridge-1-build" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.272054 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/d55d4558-c971-4a40-b1cb-41389cbf11c3-container-storage-run\") pod \"sg-bridge-1-build\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " pod="service-telemetry/sg-bridge-1-build" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.272079 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/d55d4558-c971-4a40-b1cb-41389cbf11c3-build-blob-cache\") pod \"sg-bridge-1-build\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " pod="service-telemetry/sg-bridge-1-build" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.272106 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d55d4558-c971-4a40-b1cb-41389cbf11c3-build-proxy-ca-bundles\") pod \"sg-bridge-1-build\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " pod="service-telemetry/sg-bridge-1-build" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.272098 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/d55d4558-c971-4a40-b1cb-41389cbf11c3-buildcachedir\") pod \"sg-bridge-1-build\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " pod="service-telemetry/sg-bridge-1-build" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.272154 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d55d4558-c971-4a40-b1cb-41389cbf11c3-build-ca-bundles\") pod \"sg-bridge-1-build\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " pod="service-telemetry/sg-bridge-1-build" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.272187 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/d55d4558-c971-4a40-b1cb-41389cbf11c3-builder-dockercfg-8h4xs-push\") pod \"sg-bridge-1-build\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " pod="service-telemetry/sg-bridge-1-build" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.272219 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/d55d4558-c971-4a40-b1cb-41389cbf11c3-builder-dockercfg-8h4xs-pull\") pod \"sg-bridge-1-build\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " pod="service-telemetry/sg-bridge-1-build" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.272246 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/d55d4558-c971-4a40-b1cb-41389cbf11c3-container-storage-root\") pod \"sg-bridge-1-build\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " pod="service-telemetry/sg-bridge-1-build" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.272307 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d55d4558-c971-4a40-b1cb-41389cbf11c3-node-pullsecrets\") pod \"sg-bridge-1-build\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " pod="service-telemetry/sg-bridge-1-build" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.272418 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d55d4558-c971-4a40-b1cb-41389cbf11c3-node-pullsecrets\") pod \"sg-bridge-1-build\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " pod="service-telemetry/sg-bridge-1-build" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.273028 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/d55d4558-c971-4a40-b1cb-41389cbf11c3-container-storage-run\") pod \"sg-bridge-1-build\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " pod="service-telemetry/sg-bridge-1-build" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.273298 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/d55d4558-c971-4a40-b1cb-41389cbf11c3-build-blob-cache\") pod \"sg-bridge-1-build\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " pod="service-telemetry/sg-bridge-1-build" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.273298 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/d55d4558-c971-4a40-b1cb-41389cbf11c3-buildworkdir\") pod \"sg-bridge-1-build\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " pod="service-telemetry/sg-bridge-1-build" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.273467 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/d55d4558-c971-4a40-b1cb-41389cbf11c3-container-storage-root\") pod \"sg-bridge-1-build\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " pod="service-telemetry/sg-bridge-1-build" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.273479 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/d55d4558-c971-4a40-b1cb-41389cbf11c3-build-system-configs\") pod \"sg-bridge-1-build\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " pod="service-telemetry/sg-bridge-1-build" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.273714 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d55d4558-c971-4a40-b1cb-41389cbf11c3-build-proxy-ca-bundles\") pod \"sg-bridge-1-build\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " pod="service-telemetry/sg-bridge-1-build" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.274244 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d55d4558-c971-4a40-b1cb-41389cbf11c3-build-ca-bundles\") pod \"sg-bridge-1-build\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " pod="service-telemetry/sg-bridge-1-build" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.283299 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/d55d4558-c971-4a40-b1cb-41389cbf11c3-builder-dockercfg-8h4xs-push\") pod \"sg-bridge-1-build\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " pod="service-telemetry/sg-bridge-1-build" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.283297 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/d55d4558-c971-4a40-b1cb-41389cbf11c3-builder-dockercfg-8h4xs-pull\") pod \"sg-bridge-1-build\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " pod="service-telemetry/sg-bridge-1-build" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.290935 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpgwg\" (UniqueName: \"kubernetes.io/projected/d55d4558-c971-4a40-b1cb-41389cbf11c3-kube-api-access-jpgwg\") pod \"sg-bridge-1-build\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " pod="service-telemetry/sg-bridge-1-build" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.370262 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-1-build" Jan 26 17:21:26 crc kubenswrapper[4856]: I0126 17:21:26.612326 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-bridge-1-build"] Jan 26 17:21:27 crc kubenswrapper[4856]: I0126 17:21:27.307345 4856 generic.go:334] "Generic (PLEG): container finished" podID="d55d4558-c971-4a40-b1cb-41389cbf11c3" containerID="37c0850dfb46215185250da7eccd6ad1561ff7e374d7cfc24a1386d6bf8bcf2f" exitCode=0 Jan 26 17:21:27 crc kubenswrapper[4856]: I0126 17:21:27.307410 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-1-build" event={"ID":"d55d4558-c971-4a40-b1cb-41389cbf11c3","Type":"ContainerDied","Data":"37c0850dfb46215185250da7eccd6ad1561ff7e374d7cfc24a1386d6bf8bcf2f"} Jan 26 17:21:27 crc kubenswrapper[4856]: I0126 17:21:27.307712 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-1-build" event={"ID":"d55d4558-c971-4a40-b1cb-41389cbf11c3","Type":"ContainerStarted","Data":"50a0301e7718c33315fdaefdd9a7a9347ba078d7d7dbb1b07dc7689e421c1452"} Jan 26 17:21:28 crc kubenswrapper[4856]: I0126 17:21:28.317613 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-1-build" event={"ID":"d55d4558-c971-4a40-b1cb-41389cbf11c3","Type":"ContainerStarted","Data":"1f422f325f8fa610d27451f9734bb53bd3fdded1c3e711a971293ebcfe442247"} Jan 26 17:21:28 crc kubenswrapper[4856]: I0126 17:21:28.354178 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/sg-bridge-1-build" podStartSLOduration=2.354127105 podStartE2EDuration="2.354127105s" podCreationTimestamp="2026-01-26 17:21:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:21:28.349561454 +0000 UTC m=+1384.302815455" watchObservedRunningTime="2026-01-26 17:21:28.354127105 +0000 UTC m=+1384.307381086" Jan 26 17:21:35 crc kubenswrapper[4856]: I0126 17:21:35.398384 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-bridge-1-build_d55d4558-c971-4a40-b1cb-41389cbf11c3/docker-build/0.log" Jan 26 17:21:35 crc kubenswrapper[4856]: I0126 17:21:35.399233 4856 generic.go:334] "Generic (PLEG): container finished" podID="d55d4558-c971-4a40-b1cb-41389cbf11c3" containerID="1f422f325f8fa610d27451f9734bb53bd3fdded1c3e711a971293ebcfe442247" exitCode=1 Jan 26 17:21:35 crc kubenswrapper[4856]: I0126 17:21:35.402995 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-1-build" event={"ID":"d55d4558-c971-4a40-b1cb-41389cbf11c3","Type":"ContainerDied","Data":"1f422f325f8fa610d27451f9734bb53bd3fdded1c3e711a971293ebcfe442247"} Jan 26 17:21:36 crc kubenswrapper[4856]: I0126 17:21:36.392911 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/sg-bridge-1-build"] Jan 26 17:21:36 crc kubenswrapper[4856]: I0126 17:21:36.608412 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-bridge-1-build_d55d4558-c971-4a40-b1cb-41389cbf11c3/docker-build/0.log" Jan 26 17:21:36 crc kubenswrapper[4856]: I0126 17:21:36.609067 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-1-build" Jan 26 17:21:36 crc kubenswrapper[4856]: I0126 17:21:36.715406 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/d55d4558-c971-4a40-b1cb-41389cbf11c3-build-system-configs\") pod \"d55d4558-c971-4a40-b1cb-41389cbf11c3\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " Jan 26 17:21:36 crc kubenswrapper[4856]: I0126 17:21:36.715774 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/d55d4558-c971-4a40-b1cb-41389cbf11c3-container-storage-run\") pod \"d55d4558-c971-4a40-b1cb-41389cbf11c3\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " Jan 26 17:21:36 crc kubenswrapper[4856]: I0126 17:21:36.715871 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/d55d4558-c971-4a40-b1cb-41389cbf11c3-builder-dockercfg-8h4xs-push\") pod \"d55d4558-c971-4a40-b1cb-41389cbf11c3\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " Jan 26 17:21:36 crc kubenswrapper[4856]: I0126 17:21:36.715976 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/d55d4558-c971-4a40-b1cb-41389cbf11c3-builder-dockercfg-8h4xs-pull\") pod \"d55d4558-c971-4a40-b1cb-41389cbf11c3\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " Jan 26 17:21:36 crc kubenswrapper[4856]: I0126 17:21:36.716053 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d55d4558-c971-4a40-b1cb-41389cbf11c3-node-pullsecrets\") pod \"d55d4558-c971-4a40-b1cb-41389cbf11c3\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " Jan 26 17:21:36 crc kubenswrapper[4856]: I0126 17:21:36.716154 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/d55d4558-c971-4a40-b1cb-41389cbf11c3-buildcachedir\") pod \"d55d4558-c971-4a40-b1cb-41389cbf11c3\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " Jan 26 17:21:36 crc kubenswrapper[4856]: I0126 17:21:36.716297 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d55d4558-c971-4a40-b1cb-41389cbf11c3-build-proxy-ca-bundles\") pod \"d55d4558-c971-4a40-b1cb-41389cbf11c3\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " Jan 26 17:21:36 crc kubenswrapper[4856]: I0126 17:21:36.716754 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jpgwg\" (UniqueName: \"kubernetes.io/projected/d55d4558-c971-4a40-b1cb-41389cbf11c3-kube-api-access-jpgwg\") pod \"d55d4558-c971-4a40-b1cb-41389cbf11c3\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " Jan 26 17:21:36 crc kubenswrapper[4856]: I0126 17:21:36.716893 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/d55d4558-c971-4a40-b1cb-41389cbf11c3-container-storage-root\") pod \"d55d4558-c971-4a40-b1cb-41389cbf11c3\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " Jan 26 17:21:36 crc kubenswrapper[4856]: I0126 17:21:36.719741 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d55d4558-c971-4a40-b1cb-41389cbf11c3-build-ca-bundles\") pod \"d55d4558-c971-4a40-b1cb-41389cbf11c3\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " Jan 26 17:21:36 crc kubenswrapper[4856]: I0126 17:21:36.719894 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/d55d4558-c971-4a40-b1cb-41389cbf11c3-build-blob-cache\") pod \"d55d4558-c971-4a40-b1cb-41389cbf11c3\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " Jan 26 17:21:36 crc kubenswrapper[4856]: I0126 17:21:36.719981 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/d55d4558-c971-4a40-b1cb-41389cbf11c3-buildworkdir\") pod \"d55d4558-c971-4a40-b1cb-41389cbf11c3\" (UID: \"d55d4558-c971-4a40-b1cb-41389cbf11c3\") " Jan 26 17:21:36 crc kubenswrapper[4856]: I0126 17:21:36.716169 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d55d4558-c971-4a40-b1cb-41389cbf11c3-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "d55d4558-c971-4a40-b1cb-41389cbf11c3" (UID: "d55d4558-c971-4a40-b1cb-41389cbf11c3"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:21:36 crc kubenswrapper[4856]: I0126 17:21:36.716241 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d55d4558-c971-4a40-b1cb-41389cbf11c3-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "d55d4558-c971-4a40-b1cb-41389cbf11c3" (UID: "d55d4558-c971-4a40-b1cb-41389cbf11c3"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:21:36 crc kubenswrapper[4856]: I0126 17:21:36.716371 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d55d4558-c971-4a40-b1cb-41389cbf11c3-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "d55d4558-c971-4a40-b1cb-41389cbf11c3" (UID: "d55d4558-c971-4a40-b1cb-41389cbf11c3"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:21:36 crc kubenswrapper[4856]: I0126 17:21:36.716694 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d55d4558-c971-4a40-b1cb-41389cbf11c3-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "d55d4558-c971-4a40-b1cb-41389cbf11c3" (UID: "d55d4558-c971-4a40-b1cb-41389cbf11c3"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:21:36 crc kubenswrapper[4856]: I0126 17:21:36.716701 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d55d4558-c971-4a40-b1cb-41389cbf11c3-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "d55d4558-c971-4a40-b1cb-41389cbf11c3" (UID: "d55d4558-c971-4a40-b1cb-41389cbf11c3"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:21:36 crc kubenswrapper[4856]: I0126 17:21:36.720411 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d55d4558-c971-4a40-b1cb-41389cbf11c3-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "d55d4558-c971-4a40-b1cb-41389cbf11c3" (UID: "d55d4558-c971-4a40-b1cb-41389cbf11c3"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:21:36 crc kubenswrapper[4856]: I0126 17:21:36.720435 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d55d4558-c971-4a40-b1cb-41389cbf11c3-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "d55d4558-c971-4a40-b1cb-41389cbf11c3" (UID: "d55d4558-c971-4a40-b1cb-41389cbf11c3"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:21:36 crc kubenswrapper[4856]: I0126 17:21:36.721153 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d55d4558-c971-4a40-b1cb-41389cbf11c3-kube-api-access-jpgwg" (OuterVolumeSpecName: "kube-api-access-jpgwg") pod "d55d4558-c971-4a40-b1cb-41389cbf11c3" (UID: "d55d4558-c971-4a40-b1cb-41389cbf11c3"). InnerVolumeSpecName "kube-api-access-jpgwg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:21:36 crc kubenswrapper[4856]: I0126 17:21:36.721231 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d55d4558-c971-4a40-b1cb-41389cbf11c3-builder-dockercfg-8h4xs-pull" (OuterVolumeSpecName: "builder-dockercfg-8h4xs-pull") pod "d55d4558-c971-4a40-b1cb-41389cbf11c3" (UID: "d55d4558-c971-4a40-b1cb-41389cbf11c3"). InnerVolumeSpecName "builder-dockercfg-8h4xs-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:21:36 crc kubenswrapper[4856]: I0126 17:21:36.721486 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d55d4558-c971-4a40-b1cb-41389cbf11c3-builder-dockercfg-8h4xs-push" (OuterVolumeSpecName: "builder-dockercfg-8h4xs-push") pod "d55d4558-c971-4a40-b1cb-41389cbf11c3" (UID: "d55d4558-c971-4a40-b1cb-41389cbf11c3"). InnerVolumeSpecName "builder-dockercfg-8h4xs-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:21:36 crc kubenswrapper[4856]: I0126 17:21:36.721691 4856 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/d55d4558-c971-4a40-b1cb-41389cbf11c3-buildcachedir\") on node \"crc\" DevicePath \"\"" Jan 26 17:21:36 crc kubenswrapper[4856]: I0126 17:21:36.721714 4856 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d55d4558-c971-4a40-b1cb-41389cbf11c3-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 26 17:21:36 crc kubenswrapper[4856]: I0126 17:21:36.721725 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jpgwg\" (UniqueName: \"kubernetes.io/projected/d55d4558-c971-4a40-b1cb-41389cbf11c3-kube-api-access-jpgwg\") on node \"crc\" DevicePath \"\"" Jan 26 17:21:36 crc kubenswrapper[4856]: I0126 17:21:36.721732 4856 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d55d4558-c971-4a40-b1cb-41389cbf11c3-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 26 17:21:36 crc kubenswrapper[4856]: I0126 17:21:36.721740 4856 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/d55d4558-c971-4a40-b1cb-41389cbf11c3-buildworkdir\") on node \"crc\" DevicePath \"\"" Jan 26 17:21:36 crc kubenswrapper[4856]: I0126 17:21:36.721748 4856 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/d55d4558-c971-4a40-b1cb-41389cbf11c3-build-system-configs\") on node \"crc\" DevicePath \"\"" Jan 26 17:21:36 crc kubenswrapper[4856]: I0126 17:21:36.721756 4856 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/d55d4558-c971-4a40-b1cb-41389cbf11c3-container-storage-run\") on node \"crc\" DevicePath \"\"" Jan 26 17:21:36 crc kubenswrapper[4856]: I0126 17:21:36.721774 4856 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/d55d4558-c971-4a40-b1cb-41389cbf11c3-builder-dockercfg-8h4xs-push\") on node \"crc\" DevicePath \"\"" Jan 26 17:21:36 crc kubenswrapper[4856]: I0126 17:21:36.721787 4856 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/d55d4558-c971-4a40-b1cb-41389cbf11c3-builder-dockercfg-8h4xs-pull\") on node \"crc\" DevicePath \"\"" Jan 26 17:21:36 crc kubenswrapper[4856]: I0126 17:21:36.721797 4856 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d55d4558-c971-4a40-b1cb-41389cbf11c3-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Jan 26 17:21:36 crc kubenswrapper[4856]: I0126 17:21:36.793540 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d55d4558-c971-4a40-b1cb-41389cbf11c3-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "d55d4558-c971-4a40-b1cb-41389cbf11c3" (UID: "d55d4558-c971-4a40-b1cb-41389cbf11c3"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:21:36 crc kubenswrapper[4856]: I0126 17:21:36.823306 4856 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/d55d4558-c971-4a40-b1cb-41389cbf11c3-build-blob-cache\") on node \"crc\" DevicePath \"\"" Jan 26 17:21:37 crc kubenswrapper[4856]: I0126 17:21:37.081066 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d55d4558-c971-4a40-b1cb-41389cbf11c3-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "d55d4558-c971-4a40-b1cb-41389cbf11c3" (UID: "d55d4558-c971-4a40-b1cb-41389cbf11c3"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:21:37 crc kubenswrapper[4856]: I0126 17:21:37.141794 4856 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/d55d4558-c971-4a40-b1cb-41389cbf11c3-container-storage-root\") on node \"crc\" DevicePath \"\"" Jan 26 17:21:37 crc kubenswrapper[4856]: I0126 17:21:37.411206 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-bridge-1-build_d55d4558-c971-4a40-b1cb-41389cbf11c3/docker-build/0.log" Jan 26 17:21:37 crc kubenswrapper[4856]: I0126 17:21:37.411794 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-1-build" event={"ID":"d55d4558-c971-4a40-b1cb-41389cbf11c3","Type":"ContainerDied","Data":"50a0301e7718c33315fdaefdd9a7a9347ba078d7d7dbb1b07dc7689e421c1452"} Jan 26 17:21:37 crc kubenswrapper[4856]: I0126 17:21:37.411843 4856 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="50a0301e7718c33315fdaefdd9a7a9347ba078d7d7dbb1b07dc7689e421c1452" Jan 26 17:21:37 crc kubenswrapper[4856]: I0126 17:21:37.411851 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-1-build" Jan 26 17:21:37 crc kubenswrapper[4856]: I0126 17:21:37.431413 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/sg-bridge-1-build"] Jan 26 17:21:37 crc kubenswrapper[4856]: I0126 17:21:37.437213 4856 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/sg-bridge-1-build"] Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.011031 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/sg-bridge-2-build"] Jan 26 17:21:38 crc kubenswrapper[4856]: E0126 17:21:38.011402 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d55d4558-c971-4a40-b1cb-41389cbf11c3" containerName="manage-dockerfile" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.011427 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="d55d4558-c971-4a40-b1cb-41389cbf11c3" containerName="manage-dockerfile" Jan 26 17:21:38 crc kubenswrapper[4856]: E0126 17:21:38.011458 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d55d4558-c971-4a40-b1cb-41389cbf11c3" containerName="docker-build" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.011467 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="d55d4558-c971-4a40-b1cb-41389cbf11c3" containerName="docker-build" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.011630 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="d55d4558-c971-4a40-b1cb-41389cbf11c3" containerName="docker-build" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.022702 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-2-build" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.027672 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-bridge-2-ca" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.027985 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-bridge-2-global-ca" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.028139 4856 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-8h4xs" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.028541 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-bridge-2-sys-config" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.039100 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-bridge-2-build"] Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.156291 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/d6179047-d35a-4cad-93c5-2ac126d36b6c-container-storage-run\") pod \"sg-bridge-2-build\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " pod="service-telemetry/sg-bridge-2-build" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.156349 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d6179047-d35a-4cad-93c5-2ac126d36b6c-build-ca-bundles\") pod \"sg-bridge-2-build\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " pod="service-telemetry/sg-bridge-2-build" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.156480 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/d6179047-d35a-4cad-93c5-2ac126d36b6c-buildcachedir\") pod \"sg-bridge-2-build\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " pod="service-telemetry/sg-bridge-2-build" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.156579 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/d6179047-d35a-4cad-93c5-2ac126d36b6c-container-storage-root\") pod \"sg-bridge-2-build\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " pod="service-telemetry/sg-bridge-2-build" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.156620 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/d6179047-d35a-4cad-93c5-2ac126d36b6c-builder-dockercfg-8h4xs-pull\") pod \"sg-bridge-2-build\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " pod="service-telemetry/sg-bridge-2-build" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.156679 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d6179047-d35a-4cad-93c5-2ac126d36b6c-node-pullsecrets\") pod \"sg-bridge-2-build\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " pod="service-telemetry/sg-bridge-2-build" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.156704 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjmqz\" (UniqueName: \"kubernetes.io/projected/d6179047-d35a-4cad-93c5-2ac126d36b6c-kube-api-access-sjmqz\") pod \"sg-bridge-2-build\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " pod="service-telemetry/sg-bridge-2-build" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.156721 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d6179047-d35a-4cad-93c5-2ac126d36b6c-build-proxy-ca-bundles\") pod \"sg-bridge-2-build\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " pod="service-telemetry/sg-bridge-2-build" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.156742 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/d6179047-d35a-4cad-93c5-2ac126d36b6c-builder-dockercfg-8h4xs-push\") pod \"sg-bridge-2-build\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " pod="service-telemetry/sg-bridge-2-build" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.156768 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/d6179047-d35a-4cad-93c5-2ac126d36b6c-build-system-configs\") pod \"sg-bridge-2-build\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " pod="service-telemetry/sg-bridge-2-build" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.156834 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/d6179047-d35a-4cad-93c5-2ac126d36b6c-buildworkdir\") pod \"sg-bridge-2-build\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " pod="service-telemetry/sg-bridge-2-build" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.156878 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/d6179047-d35a-4cad-93c5-2ac126d36b6c-build-blob-cache\") pod \"sg-bridge-2-build\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " pod="service-telemetry/sg-bridge-2-build" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.258093 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d6179047-d35a-4cad-93c5-2ac126d36b6c-node-pullsecrets\") pod \"sg-bridge-2-build\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " pod="service-telemetry/sg-bridge-2-build" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.258160 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjmqz\" (UniqueName: \"kubernetes.io/projected/d6179047-d35a-4cad-93c5-2ac126d36b6c-kube-api-access-sjmqz\") pod \"sg-bridge-2-build\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " pod="service-telemetry/sg-bridge-2-build" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.258186 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d6179047-d35a-4cad-93c5-2ac126d36b6c-build-proxy-ca-bundles\") pod \"sg-bridge-2-build\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " pod="service-telemetry/sg-bridge-2-build" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.258219 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/d6179047-d35a-4cad-93c5-2ac126d36b6c-builder-dockercfg-8h4xs-push\") pod \"sg-bridge-2-build\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " pod="service-telemetry/sg-bridge-2-build" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.258246 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/d6179047-d35a-4cad-93c5-2ac126d36b6c-build-system-configs\") pod \"sg-bridge-2-build\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " pod="service-telemetry/sg-bridge-2-build" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.258291 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/d6179047-d35a-4cad-93c5-2ac126d36b6c-buildworkdir\") pod \"sg-bridge-2-build\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " pod="service-telemetry/sg-bridge-2-build" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.258315 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/d6179047-d35a-4cad-93c5-2ac126d36b6c-build-blob-cache\") pod \"sg-bridge-2-build\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " pod="service-telemetry/sg-bridge-2-build" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.258351 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/d6179047-d35a-4cad-93c5-2ac126d36b6c-container-storage-run\") pod \"sg-bridge-2-build\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " pod="service-telemetry/sg-bridge-2-build" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.258379 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d6179047-d35a-4cad-93c5-2ac126d36b6c-build-ca-bundles\") pod \"sg-bridge-2-build\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " pod="service-telemetry/sg-bridge-2-build" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.258402 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/d6179047-d35a-4cad-93c5-2ac126d36b6c-buildcachedir\") pod \"sg-bridge-2-build\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " pod="service-telemetry/sg-bridge-2-build" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.258440 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/d6179047-d35a-4cad-93c5-2ac126d36b6c-container-storage-root\") pod \"sg-bridge-2-build\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " pod="service-telemetry/sg-bridge-2-build" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.258487 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/d6179047-d35a-4cad-93c5-2ac126d36b6c-builder-dockercfg-8h4xs-pull\") pod \"sg-bridge-2-build\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " pod="service-telemetry/sg-bridge-2-build" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.258850 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/d6179047-d35a-4cad-93c5-2ac126d36b6c-buildcachedir\") pod \"sg-bridge-2-build\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " pod="service-telemetry/sg-bridge-2-build" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.258910 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/d6179047-d35a-4cad-93c5-2ac126d36b6c-buildworkdir\") pod \"sg-bridge-2-build\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " pod="service-telemetry/sg-bridge-2-build" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.259116 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/d6179047-d35a-4cad-93c5-2ac126d36b6c-build-blob-cache\") pod \"sg-bridge-2-build\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " pod="service-telemetry/sg-bridge-2-build" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.259228 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/d6179047-d35a-4cad-93c5-2ac126d36b6c-container-storage-root\") pod \"sg-bridge-2-build\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " pod="service-telemetry/sg-bridge-2-build" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.259246 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d6179047-d35a-4cad-93c5-2ac126d36b6c-node-pullsecrets\") pod \"sg-bridge-2-build\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " pod="service-telemetry/sg-bridge-2-build" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.259276 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d6179047-d35a-4cad-93c5-2ac126d36b6c-build-proxy-ca-bundles\") pod \"sg-bridge-2-build\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " pod="service-telemetry/sg-bridge-2-build" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.259281 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/d6179047-d35a-4cad-93c5-2ac126d36b6c-container-storage-run\") pod \"sg-bridge-2-build\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " pod="service-telemetry/sg-bridge-2-build" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.259793 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d6179047-d35a-4cad-93c5-2ac126d36b6c-build-ca-bundles\") pod \"sg-bridge-2-build\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " pod="service-telemetry/sg-bridge-2-build" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.259818 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/d6179047-d35a-4cad-93c5-2ac126d36b6c-build-system-configs\") pod \"sg-bridge-2-build\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " pod="service-telemetry/sg-bridge-2-build" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.264623 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/d6179047-d35a-4cad-93c5-2ac126d36b6c-builder-dockercfg-8h4xs-push\") pod \"sg-bridge-2-build\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " pod="service-telemetry/sg-bridge-2-build" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.266741 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/d6179047-d35a-4cad-93c5-2ac126d36b6c-builder-dockercfg-8h4xs-pull\") pod \"sg-bridge-2-build\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " pod="service-telemetry/sg-bridge-2-build" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.275157 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjmqz\" (UniqueName: \"kubernetes.io/projected/d6179047-d35a-4cad-93c5-2ac126d36b6c-kube-api-access-sjmqz\") pod \"sg-bridge-2-build\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " pod="service-telemetry/sg-bridge-2-build" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.391952 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-2-build" Jan 26 17:21:38 crc kubenswrapper[4856]: I0126 17:21:38.587113 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-bridge-2-build"] Jan 26 17:21:39 crc kubenswrapper[4856]: I0126 17:21:39.403288 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d55d4558-c971-4a40-b1cb-41389cbf11c3" path="/var/lib/kubelet/pods/d55d4558-c971-4a40-b1cb-41389cbf11c3/volumes" Jan 26 17:21:39 crc kubenswrapper[4856]: I0126 17:21:39.424300 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"d6179047-d35a-4cad-93c5-2ac126d36b6c","Type":"ContainerStarted","Data":"45628591f87569fa9e26c4179bfa65038a601dfa6e79bf1c2dc16826001a2368"} Jan 26 17:21:39 crc kubenswrapper[4856]: I0126 17:21:39.424357 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"d6179047-d35a-4cad-93c5-2ac126d36b6c","Type":"ContainerStarted","Data":"6a8b6988fa5a1ca6aa116299ba279d842c443263b949816e3eeba14922cdf1f8"} Jan 26 17:21:40 crc kubenswrapper[4856]: I0126 17:21:40.515123 4856 generic.go:334] "Generic (PLEG): container finished" podID="d6179047-d35a-4cad-93c5-2ac126d36b6c" containerID="45628591f87569fa9e26c4179bfa65038a601dfa6e79bf1c2dc16826001a2368" exitCode=0 Jan 26 17:21:40 crc kubenswrapper[4856]: I0126 17:21:40.515218 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"d6179047-d35a-4cad-93c5-2ac126d36b6c","Type":"ContainerDied","Data":"45628591f87569fa9e26c4179bfa65038a601dfa6e79bf1c2dc16826001a2368"} Jan 26 17:21:41 crc kubenswrapper[4856]: I0126 17:21:41.538310 4856 generic.go:334] "Generic (PLEG): container finished" podID="d6179047-d35a-4cad-93c5-2ac126d36b6c" containerID="40e8aa2782ef0e0d6ad31061c662a5d62042428d1e0ef84a274003729ce0e44b" exitCode=0 Jan 26 17:21:41 crc kubenswrapper[4856]: I0126 17:21:41.538359 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"d6179047-d35a-4cad-93c5-2ac126d36b6c","Type":"ContainerDied","Data":"40e8aa2782ef0e0d6ad31061c662a5d62042428d1e0ef84a274003729ce0e44b"} Jan 26 17:21:41 crc kubenswrapper[4856]: I0126 17:21:41.584866 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-bridge-2-build_d6179047-d35a-4cad-93c5-2ac126d36b6c/manage-dockerfile/0.log" Jan 26 17:21:42 crc kubenswrapper[4856]: I0126 17:21:42.547586 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"d6179047-d35a-4cad-93c5-2ac126d36b6c","Type":"ContainerStarted","Data":"0708c3ea94fa93ca300b24c4b8401e5affed40b2e66cd659a6c94ff571bc0799"} Jan 26 17:21:42 crc kubenswrapper[4856]: I0126 17:21:42.577021 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/sg-bridge-2-build" podStartSLOduration=5.576984808 podStartE2EDuration="5.576984808s" podCreationTimestamp="2026-01-26 17:21:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:21:42.574421494 +0000 UTC m=+1398.527675475" watchObservedRunningTime="2026-01-26 17:21:42.576984808 +0000 UTC m=+1398.530238809" Jan 26 17:22:25 crc kubenswrapper[4856]: E0126 17:22:25.375388 4856 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6179047_d35a_4cad_93c5_2ac126d36b6c.slice/buildah-buildah1089185293\": RecentStats: unable to find data in memory cache]" Jan 26 17:22:26 crc kubenswrapper[4856]: I0126 17:22:26.939656 4856 patch_prober.go:28] interesting pod/machine-config-daemon-xm9cq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:22:26 crc kubenswrapper[4856]: I0126 17:22:26.940310 4856 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:22:31 crc kubenswrapper[4856]: I0126 17:22:31.940825 4856 generic.go:334] "Generic (PLEG): container finished" podID="d6179047-d35a-4cad-93c5-2ac126d36b6c" containerID="0708c3ea94fa93ca300b24c4b8401e5affed40b2e66cd659a6c94ff571bc0799" exitCode=0 Jan 26 17:22:31 crc kubenswrapper[4856]: I0126 17:22:31.940900 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"d6179047-d35a-4cad-93c5-2ac126d36b6c","Type":"ContainerDied","Data":"0708c3ea94fa93ca300b24c4b8401e5affed40b2e66cd659a6c94ff571bc0799"} Jan 26 17:22:33 crc kubenswrapper[4856]: I0126 17:22:33.194187 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-2-build" Jan 26 17:22:33 crc kubenswrapper[4856]: I0126 17:22:33.304774 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/d6179047-d35a-4cad-93c5-2ac126d36b6c-build-blob-cache\") pod \"d6179047-d35a-4cad-93c5-2ac126d36b6c\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " Jan 26 17:22:33 crc kubenswrapper[4856]: I0126 17:22:33.304838 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/d6179047-d35a-4cad-93c5-2ac126d36b6c-builder-dockercfg-8h4xs-push\") pod \"d6179047-d35a-4cad-93c5-2ac126d36b6c\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " Jan 26 17:22:33 crc kubenswrapper[4856]: I0126 17:22:33.304884 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d6179047-d35a-4cad-93c5-2ac126d36b6c-build-ca-bundles\") pod \"d6179047-d35a-4cad-93c5-2ac126d36b6c\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " Jan 26 17:22:33 crc kubenswrapper[4856]: I0126 17:22:33.304938 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/d6179047-d35a-4cad-93c5-2ac126d36b6c-container-storage-run\") pod \"d6179047-d35a-4cad-93c5-2ac126d36b6c\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " Jan 26 17:22:33 crc kubenswrapper[4856]: I0126 17:22:33.305837 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/d6179047-d35a-4cad-93c5-2ac126d36b6c-build-system-configs\") pod \"d6179047-d35a-4cad-93c5-2ac126d36b6c\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " Jan 26 17:22:33 crc kubenswrapper[4856]: I0126 17:22:33.305876 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/d6179047-d35a-4cad-93c5-2ac126d36b6c-buildworkdir\") pod \"d6179047-d35a-4cad-93c5-2ac126d36b6c\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " Jan 26 17:22:33 crc kubenswrapper[4856]: I0126 17:22:33.305899 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d6179047-d35a-4cad-93c5-2ac126d36b6c-build-proxy-ca-bundles\") pod \"d6179047-d35a-4cad-93c5-2ac126d36b6c\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " Jan 26 17:22:33 crc kubenswrapper[4856]: I0126 17:22:33.305923 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/d6179047-d35a-4cad-93c5-2ac126d36b6c-buildcachedir\") pod \"d6179047-d35a-4cad-93c5-2ac126d36b6c\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " Jan 26 17:22:33 crc kubenswrapper[4856]: I0126 17:22:33.305998 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjmqz\" (UniqueName: \"kubernetes.io/projected/d6179047-d35a-4cad-93c5-2ac126d36b6c-kube-api-access-sjmqz\") pod \"d6179047-d35a-4cad-93c5-2ac126d36b6c\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " Jan 26 17:22:33 crc kubenswrapper[4856]: I0126 17:22:33.306029 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/d6179047-d35a-4cad-93c5-2ac126d36b6c-container-storage-root\") pod \"d6179047-d35a-4cad-93c5-2ac126d36b6c\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " Jan 26 17:22:33 crc kubenswrapper[4856]: I0126 17:22:33.306052 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/d6179047-d35a-4cad-93c5-2ac126d36b6c-builder-dockercfg-8h4xs-pull\") pod \"d6179047-d35a-4cad-93c5-2ac126d36b6c\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " Jan 26 17:22:33 crc kubenswrapper[4856]: I0126 17:22:33.306078 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d6179047-d35a-4cad-93c5-2ac126d36b6c-node-pullsecrets\") pod \"d6179047-d35a-4cad-93c5-2ac126d36b6c\" (UID: \"d6179047-d35a-4cad-93c5-2ac126d36b6c\") " Jan 26 17:22:33 crc kubenswrapper[4856]: I0126 17:22:33.306227 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6179047-d35a-4cad-93c5-2ac126d36b6c-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "d6179047-d35a-4cad-93c5-2ac126d36b6c" (UID: "d6179047-d35a-4cad-93c5-2ac126d36b6c"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:22:33 crc kubenswrapper[4856]: I0126 17:22:33.306345 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d6179047-d35a-4cad-93c5-2ac126d36b6c-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "d6179047-d35a-4cad-93c5-2ac126d36b6c" (UID: "d6179047-d35a-4cad-93c5-2ac126d36b6c"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:22:33 crc kubenswrapper[4856]: I0126 17:22:33.306481 4856 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/d6179047-d35a-4cad-93c5-2ac126d36b6c-buildcachedir\") on node \"crc\" DevicePath \"\"" Jan 26 17:22:33 crc kubenswrapper[4856]: I0126 17:22:33.306586 4856 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d6179047-d35a-4cad-93c5-2ac126d36b6c-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Jan 26 17:22:33 crc kubenswrapper[4856]: I0126 17:22:33.306866 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6179047-d35a-4cad-93c5-2ac126d36b6c-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "d6179047-d35a-4cad-93c5-2ac126d36b6c" (UID: "d6179047-d35a-4cad-93c5-2ac126d36b6c"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:22:33 crc kubenswrapper[4856]: I0126 17:22:33.307089 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6179047-d35a-4cad-93c5-2ac126d36b6c-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "d6179047-d35a-4cad-93c5-2ac126d36b6c" (UID: "d6179047-d35a-4cad-93c5-2ac126d36b6c"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:22:33 crc kubenswrapper[4856]: I0126 17:22:33.307265 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6179047-d35a-4cad-93c5-2ac126d36b6c-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "d6179047-d35a-4cad-93c5-2ac126d36b6c" (UID: "d6179047-d35a-4cad-93c5-2ac126d36b6c"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:22:33 crc kubenswrapper[4856]: I0126 17:22:33.307435 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6179047-d35a-4cad-93c5-2ac126d36b6c-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "d6179047-d35a-4cad-93c5-2ac126d36b6c" (UID: "d6179047-d35a-4cad-93c5-2ac126d36b6c"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:22:33 crc kubenswrapper[4856]: I0126 17:22:33.308508 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6179047-d35a-4cad-93c5-2ac126d36b6c-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "d6179047-d35a-4cad-93c5-2ac126d36b6c" (UID: "d6179047-d35a-4cad-93c5-2ac126d36b6c"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:22:33 crc kubenswrapper[4856]: I0126 17:22:33.316574 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6179047-d35a-4cad-93c5-2ac126d36b6c-builder-dockercfg-8h4xs-push" (OuterVolumeSpecName: "builder-dockercfg-8h4xs-push") pod "d6179047-d35a-4cad-93c5-2ac126d36b6c" (UID: "d6179047-d35a-4cad-93c5-2ac126d36b6c"). InnerVolumeSpecName "builder-dockercfg-8h4xs-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:22:33 crc kubenswrapper[4856]: I0126 17:22:33.333934 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6179047-d35a-4cad-93c5-2ac126d36b6c-kube-api-access-sjmqz" (OuterVolumeSpecName: "kube-api-access-sjmqz") pod "d6179047-d35a-4cad-93c5-2ac126d36b6c" (UID: "d6179047-d35a-4cad-93c5-2ac126d36b6c"). InnerVolumeSpecName "kube-api-access-sjmqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:22:33 crc kubenswrapper[4856]: I0126 17:22:33.341731 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6179047-d35a-4cad-93c5-2ac126d36b6c-builder-dockercfg-8h4xs-pull" (OuterVolumeSpecName: "builder-dockercfg-8h4xs-pull") pod "d6179047-d35a-4cad-93c5-2ac126d36b6c" (UID: "d6179047-d35a-4cad-93c5-2ac126d36b6c"). InnerVolumeSpecName "builder-dockercfg-8h4xs-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:22:33 crc kubenswrapper[4856]: I0126 17:22:33.409391 4856 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/d6179047-d35a-4cad-93c5-2ac126d36b6c-builder-dockercfg-8h4xs-push\") on node \"crc\" DevicePath \"\"" Jan 26 17:22:33 crc kubenswrapper[4856]: I0126 17:22:33.409450 4856 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d6179047-d35a-4cad-93c5-2ac126d36b6c-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 26 17:22:33 crc kubenswrapper[4856]: I0126 17:22:33.409462 4856 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/d6179047-d35a-4cad-93c5-2ac126d36b6c-container-storage-run\") on node \"crc\" DevicePath \"\"" Jan 26 17:22:33 crc kubenswrapper[4856]: I0126 17:22:33.409472 4856 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/d6179047-d35a-4cad-93c5-2ac126d36b6c-build-system-configs\") on node \"crc\" DevicePath \"\"" Jan 26 17:22:33 crc kubenswrapper[4856]: I0126 17:22:33.409484 4856 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/d6179047-d35a-4cad-93c5-2ac126d36b6c-buildworkdir\") on node \"crc\" DevicePath \"\"" Jan 26 17:22:33 crc kubenswrapper[4856]: I0126 17:22:33.409494 4856 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d6179047-d35a-4cad-93c5-2ac126d36b6c-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 26 17:22:33 crc kubenswrapper[4856]: I0126 17:22:33.409502 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjmqz\" (UniqueName: \"kubernetes.io/projected/d6179047-d35a-4cad-93c5-2ac126d36b6c-kube-api-access-sjmqz\") on node \"crc\" DevicePath \"\"" Jan 26 17:22:33 crc kubenswrapper[4856]: I0126 17:22:33.409510 4856 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/d6179047-d35a-4cad-93c5-2ac126d36b6c-builder-dockercfg-8h4xs-pull\") on node \"crc\" DevicePath \"\"" Jan 26 17:22:33 crc kubenswrapper[4856]: I0126 17:22:33.559631 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6179047-d35a-4cad-93c5-2ac126d36b6c-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "d6179047-d35a-4cad-93c5-2ac126d36b6c" (UID: "d6179047-d35a-4cad-93c5-2ac126d36b6c"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:22:33 crc kubenswrapper[4856]: I0126 17:22:33.612081 4856 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/d6179047-d35a-4cad-93c5-2ac126d36b6c-build-blob-cache\") on node \"crc\" DevicePath \"\"" Jan 26 17:22:33 crc kubenswrapper[4856]: I0126 17:22:33.958490 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"d6179047-d35a-4cad-93c5-2ac126d36b6c","Type":"ContainerDied","Data":"6a8b6988fa5a1ca6aa116299ba279d842c443263b949816e3eeba14922cdf1f8"} Jan 26 17:22:33 crc kubenswrapper[4856]: I0126 17:22:33.958616 4856 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6a8b6988fa5a1ca6aa116299ba279d842c443263b949816e3eeba14922cdf1f8" Jan 26 17:22:33 crc kubenswrapper[4856]: I0126 17:22:33.958628 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-2-build" Jan 26 17:22:34 crc kubenswrapper[4856]: I0126 17:22:34.044579 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6179047-d35a-4cad-93c5-2ac126d36b6c-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "d6179047-d35a-4cad-93c5-2ac126d36b6c" (UID: "d6179047-d35a-4cad-93c5-2ac126d36b6c"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:22:34 crc kubenswrapper[4856]: I0126 17:22:34.118083 4856 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/d6179047-d35a-4cad-93c5-2ac126d36b6c-container-storage-root\") on node \"crc\" DevicePath \"\"" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.130033 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/prometheus-webhook-snmp-1-build"] Jan 26 17:22:38 crc kubenswrapper[4856]: E0126 17:22:38.130646 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6179047-d35a-4cad-93c5-2ac126d36b6c" containerName="git-clone" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.130663 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6179047-d35a-4cad-93c5-2ac126d36b6c" containerName="git-clone" Jan 26 17:22:38 crc kubenswrapper[4856]: E0126 17:22:38.130673 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6179047-d35a-4cad-93c5-2ac126d36b6c" containerName="manage-dockerfile" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.130680 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6179047-d35a-4cad-93c5-2ac126d36b6c" containerName="manage-dockerfile" Jan 26 17:22:38 crc kubenswrapper[4856]: E0126 17:22:38.130697 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6179047-d35a-4cad-93c5-2ac126d36b6c" containerName="docker-build" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.130709 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6179047-d35a-4cad-93c5-2ac126d36b6c" containerName="docker-build" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.130890 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6179047-d35a-4cad-93c5-2ac126d36b6c" containerName="docker-build" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.131646 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.133672 4856 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-8h4xs" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.134105 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-1-global-ca" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.134188 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-1-ca" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.134314 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-1-sys-config" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.150763 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-1-build"] Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.174091 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.174144 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-node-pullsecrets\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.174219 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-builder-dockercfg-8h4xs-push\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.174281 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-container-storage-root\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.174316 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-build-system-configs\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.174361 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-buildcachedir\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.174403 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-container-storage-run\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.174440 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-build-ca-bundles\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.174477 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-builder-dockercfg-8h4xs-pull\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.174496 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-build-blob-cache\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.174512 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-buildworkdir\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.174550 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9fg2\" (UniqueName: \"kubernetes.io/projected/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-kube-api-access-q9fg2\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.275538 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-builder-dockercfg-8h4xs-pull\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.275583 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-build-blob-cache\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.275610 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-buildworkdir\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.275635 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9fg2\" (UniqueName: \"kubernetes.io/projected/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-kube-api-access-q9fg2\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.275655 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.275675 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-node-pullsecrets\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.275695 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-builder-dockercfg-8h4xs-push\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.275717 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-container-storage-root\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.275737 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-build-system-configs\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.275758 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-buildcachedir\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.275775 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-container-storage-run\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.275793 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-build-ca-bundles\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.276484 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-build-ca-bundles\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.276727 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-build-blob-cache\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.276904 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-buildworkdir\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.277610 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.277661 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-node-pullsecrets\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.278080 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-buildcachedir\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.278227 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-build-system-configs\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.278361 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-container-storage-root\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.278569 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-container-storage-run\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.281031 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-builder-dockercfg-8h4xs-push\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.281307 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-builder-dockercfg-8h4xs-pull\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.296072 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9fg2\" (UniqueName: \"kubernetes.io/projected/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-kube-api-access-q9fg2\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.447130 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.674519 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-1-build"] Jan 26 17:22:38 crc kubenswrapper[4856]: I0126 17:22:38.993484 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-1-build" event={"ID":"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0","Type":"ContainerStarted","Data":"3a3db54af766ff6e3ccc317cea1e0c54cb4eb5de533c2e962ee1bc8a2be3b885"} Jan 26 17:22:40 crc kubenswrapper[4856]: I0126 17:22:40.002797 4856 generic.go:334] "Generic (PLEG): container finished" podID="46e64022-34f9-4df3-a5aa-a8b9f20a4cb0" containerID="403583ea8675bdc45ae90876cfddf217f8d8287e642b81234cb02aced617aab6" exitCode=0 Jan 26 17:22:40 crc kubenswrapper[4856]: I0126 17:22:40.002893 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-1-build" event={"ID":"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0","Type":"ContainerDied","Data":"403583ea8675bdc45ae90876cfddf217f8d8287e642b81234cb02aced617aab6"} Jan 26 17:22:41 crc kubenswrapper[4856]: I0126 17:22:41.011413 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-1-build" event={"ID":"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0","Type":"ContainerStarted","Data":"e5f538bd0f87a5c33bc9d1f0b968a1b8c16df013efd23455a52c388972a23731"} Jan 26 17:22:41 crc kubenswrapper[4856]: I0126 17:22:41.035591 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/prometheus-webhook-snmp-1-build" podStartSLOduration=3.035569627 podStartE2EDuration="3.035569627s" podCreationTimestamp="2026-01-26 17:22:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:22:41.031438719 +0000 UTC m=+1456.984692720" watchObservedRunningTime="2026-01-26 17:22:41.035569627 +0000 UTC m=+1456.988823628" Jan 26 17:22:49 crc kubenswrapper[4856]: I0126 17:22:49.239173 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-1-build"] Jan 26 17:22:49 crc kubenswrapper[4856]: I0126 17:22:49.240008 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/prometheus-webhook-snmp-1-build" podUID="46e64022-34f9-4df3-a5aa-a8b9f20a4cb0" containerName="docker-build" containerID="cri-o://e5f538bd0f87a5c33bc9d1f0b968a1b8c16df013efd23455a52c388972a23731" gracePeriod=30 Jan 26 17:22:50 crc kubenswrapper[4856]: I0126 17:22:50.075336 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-webhook-snmp-1-build_46e64022-34f9-4df3-a5aa-a8b9f20a4cb0/docker-build/0.log" Jan 26 17:22:50 crc kubenswrapper[4856]: I0126 17:22:50.076174 4856 generic.go:334] "Generic (PLEG): container finished" podID="46e64022-34f9-4df3-a5aa-a8b9f20a4cb0" containerID="e5f538bd0f87a5c33bc9d1f0b968a1b8c16df013efd23455a52c388972a23731" exitCode=1 Jan 26 17:22:50 crc kubenswrapper[4856]: I0126 17:22:50.076224 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-1-build" event={"ID":"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0","Type":"ContainerDied","Data":"e5f538bd0f87a5c33bc9d1f0b968a1b8c16df013efd23455a52c388972a23731"} Jan 26 17:22:50 crc kubenswrapper[4856]: I0126 17:22:50.140878 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-webhook-snmp-1-build_46e64022-34f9-4df3-a5aa-a8b9f20a4cb0/docker-build/0.log" Jan 26 17:22:50 crc kubenswrapper[4856]: I0126 17:22:50.141739 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 26 17:22:50 crc kubenswrapper[4856]: I0126 17:22:50.248591 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-node-pullsecrets\") pod \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " Jan 26 17:22:50 crc kubenswrapper[4856]: I0126 17:22:50.248637 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-buildcachedir\") pod \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " Jan 26 17:22:50 crc kubenswrapper[4856]: I0126 17:22:50.248681 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-container-storage-root\") pod \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " Jan 26 17:22:50 crc kubenswrapper[4856]: I0126 17:22:50.248745 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-build-blob-cache\") pod \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " Jan 26 17:22:50 crc kubenswrapper[4856]: I0126 17:22:50.248743 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "46e64022-34f9-4df3-a5aa-a8b9f20a4cb0" (UID: "46e64022-34f9-4df3-a5aa-a8b9f20a4cb0"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:22:50 crc kubenswrapper[4856]: I0126 17:22:50.248781 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9fg2\" (UniqueName: \"kubernetes.io/projected/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-kube-api-access-q9fg2\") pod \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " Jan 26 17:22:50 crc kubenswrapper[4856]: I0126 17:22:50.248816 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-builder-dockercfg-8h4xs-push\") pod \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " Jan 26 17:22:50 crc kubenswrapper[4856]: I0126 17:22:50.248842 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-buildworkdir\") pod \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " Jan 26 17:22:50 crc kubenswrapper[4856]: I0126 17:22:50.248873 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-build-proxy-ca-bundles\") pod \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " Jan 26 17:22:50 crc kubenswrapper[4856]: I0126 17:22:50.248895 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-builder-dockercfg-8h4xs-pull\") pod \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " Jan 26 17:22:50 crc kubenswrapper[4856]: I0126 17:22:50.248904 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "46e64022-34f9-4df3-a5aa-a8b9f20a4cb0" (UID: "46e64022-34f9-4df3-a5aa-a8b9f20a4cb0"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:22:50 crc kubenswrapper[4856]: I0126 17:22:50.248920 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-build-system-configs\") pod \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " Jan 26 17:22:50 crc kubenswrapper[4856]: I0126 17:22:50.248972 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-container-storage-run\") pod \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " Jan 26 17:22:50 crc kubenswrapper[4856]: I0126 17:22:50.248997 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-build-ca-bundles\") pod \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\" (UID: \"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0\") " Jan 26 17:22:50 crc kubenswrapper[4856]: I0126 17:22:50.249177 4856 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Jan 26 17:22:50 crc kubenswrapper[4856]: I0126 17:22:50.249192 4856 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-buildcachedir\") on node \"crc\" DevicePath \"\"" Jan 26 17:22:50 crc kubenswrapper[4856]: I0126 17:22:50.249413 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "46e64022-34f9-4df3-a5aa-a8b9f20a4cb0" (UID: "46e64022-34f9-4df3-a5aa-a8b9f20a4cb0"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:22:50 crc kubenswrapper[4856]: I0126 17:22:50.249902 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "46e64022-34f9-4df3-a5aa-a8b9f20a4cb0" (UID: "46e64022-34f9-4df3-a5aa-a8b9f20a4cb0"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:22:50 crc kubenswrapper[4856]: I0126 17:22:50.250328 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "46e64022-34f9-4df3-a5aa-a8b9f20a4cb0" (UID: "46e64022-34f9-4df3-a5aa-a8b9f20a4cb0"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:22:50 crc kubenswrapper[4856]: I0126 17:22:50.250880 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "46e64022-34f9-4df3-a5aa-a8b9f20a4cb0" (UID: "46e64022-34f9-4df3-a5aa-a8b9f20a4cb0"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:22:50 crc kubenswrapper[4856]: I0126 17:22:50.251312 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "46e64022-34f9-4df3-a5aa-a8b9f20a4cb0" (UID: "46e64022-34f9-4df3-a5aa-a8b9f20a4cb0"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:22:50 crc kubenswrapper[4856]: I0126 17:22:50.261270 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-builder-dockercfg-8h4xs-push" (OuterVolumeSpecName: "builder-dockercfg-8h4xs-push") pod "46e64022-34f9-4df3-a5aa-a8b9f20a4cb0" (UID: "46e64022-34f9-4df3-a5aa-a8b9f20a4cb0"). InnerVolumeSpecName "builder-dockercfg-8h4xs-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:22:50 crc kubenswrapper[4856]: I0126 17:22:50.261844 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-kube-api-access-q9fg2" (OuterVolumeSpecName: "kube-api-access-q9fg2") pod "46e64022-34f9-4df3-a5aa-a8b9f20a4cb0" (UID: "46e64022-34f9-4df3-a5aa-a8b9f20a4cb0"). InnerVolumeSpecName "kube-api-access-q9fg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:22:50 crc kubenswrapper[4856]: I0126 17:22:50.271574 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-builder-dockercfg-8h4xs-pull" (OuterVolumeSpecName: "builder-dockercfg-8h4xs-pull") pod "46e64022-34f9-4df3-a5aa-a8b9f20a4cb0" (UID: "46e64022-34f9-4df3-a5aa-a8b9f20a4cb0"). InnerVolumeSpecName "builder-dockercfg-8h4xs-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:22:50 crc kubenswrapper[4856]: I0126 17:22:50.321915 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "46e64022-34f9-4df3-a5aa-a8b9f20a4cb0" (UID: "46e64022-34f9-4df3-a5aa-a8b9f20a4cb0"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:22:50 crc kubenswrapper[4856]: I0126 17:22:50.349942 4856 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-builder-dockercfg-8h4xs-push\") on node \"crc\" DevicePath \"\"" Jan 26 17:22:50 crc kubenswrapper[4856]: I0126 17:22:50.349976 4856 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-buildworkdir\") on node \"crc\" DevicePath \"\"" Jan 26 17:22:50 crc kubenswrapper[4856]: I0126 17:22:50.349987 4856 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 26 17:22:50 crc kubenswrapper[4856]: I0126 17:22:50.349996 4856 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-builder-dockercfg-8h4xs-pull\") on node \"crc\" DevicePath \"\"" Jan 26 17:22:50 crc kubenswrapper[4856]: I0126 17:22:50.350004 4856 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-build-system-configs\") on node \"crc\" DevicePath \"\"" Jan 26 17:22:50 crc kubenswrapper[4856]: I0126 17:22:50.350012 4856 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-container-storage-run\") on node \"crc\" DevicePath \"\"" Jan 26 17:22:50 crc kubenswrapper[4856]: I0126 17:22:50.350020 4856 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 26 17:22:50 crc kubenswrapper[4856]: I0126 17:22:50.350028 4856 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-build-blob-cache\") on node \"crc\" DevicePath \"\"" Jan 26 17:22:50 crc kubenswrapper[4856]: I0126 17:22:50.350036 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9fg2\" (UniqueName: \"kubernetes.io/projected/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-kube-api-access-q9fg2\") on node \"crc\" DevicePath \"\"" Jan 26 17:22:50 crc kubenswrapper[4856]: I0126 17:22:50.655088 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "46e64022-34f9-4df3-a5aa-a8b9f20a4cb0" (UID: "46e64022-34f9-4df3-a5aa-a8b9f20a4cb0"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:22:50 crc kubenswrapper[4856]: I0126 17:22:50.761167 4856 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0-container-storage-root\") on node \"crc\" DevicePath \"\"" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.084291 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-webhook-snmp-1-build_46e64022-34f9-4df3-a5aa-a8b9f20a4cb0/docker-build/0.log" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.084742 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-1-build" event={"ID":"46e64022-34f9-4df3-a5aa-a8b9f20a4cb0","Type":"ContainerDied","Data":"3a3db54af766ff6e3ccc317cea1e0c54cb4eb5de533c2e962ee1bc8a2be3b885"} Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.084817 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-1-build" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.084841 4856 scope.go:117] "RemoveContainer" containerID="e5f538bd0f87a5c33bc9d1f0b968a1b8c16df013efd23455a52c388972a23731" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.114678 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-1-build"] Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.117674 4856 scope.go:117] "RemoveContainer" containerID="403583ea8675bdc45ae90876cfddf217f8d8287e642b81234cb02aced617aab6" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.122380 4856 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-1-build"] Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.378102 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/prometheus-webhook-snmp-2-build"] Jan 26 17:22:51 crc kubenswrapper[4856]: E0126 17:22:51.378800 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46e64022-34f9-4df3-a5aa-a8b9f20a4cb0" containerName="docker-build" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.378819 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="46e64022-34f9-4df3-a5aa-a8b9f20a4cb0" containerName="docker-build" Jan 26 17:22:51 crc kubenswrapper[4856]: E0126 17:22:51.378847 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46e64022-34f9-4df3-a5aa-a8b9f20a4cb0" containerName="manage-dockerfile" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.378858 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="46e64022-34f9-4df3-a5aa-a8b9f20a4cb0" containerName="manage-dockerfile" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.379003 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="46e64022-34f9-4df3-a5aa-a8b9f20a4cb0" containerName="docker-build" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.380105 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.383166 4856 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-8h4xs" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.383760 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-2-sys-config" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.384612 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-2-ca" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.384973 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-2-global-ca" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.406289 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46e64022-34f9-4df3-a5aa-a8b9f20a4cb0" path="/var/lib/kubelet/pods/46e64022-34f9-4df3-a5aa-a8b9f20a4cb0/volumes" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.407265 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-2-build"] Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.575135 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/605203af-fcdf-42c0-a66f-5c412f8e7770-container-storage-root\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.575204 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/605203af-fcdf-42c0-a66f-5c412f8e7770-node-pullsecrets\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.575236 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/605203af-fcdf-42c0-a66f-5c412f8e7770-buildcachedir\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.575258 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/605203af-fcdf-42c0-a66f-5c412f8e7770-build-system-configs\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.575330 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/605203af-fcdf-42c0-a66f-5c412f8e7770-buildworkdir\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.575358 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/605203af-fcdf-42c0-a66f-5c412f8e7770-build-ca-bundles\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.575465 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/605203af-fcdf-42c0-a66f-5c412f8e7770-builder-dockercfg-8h4xs-push\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.575559 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/605203af-fcdf-42c0-a66f-5c412f8e7770-container-storage-run\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.575586 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/605203af-fcdf-42c0-a66f-5c412f8e7770-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.575622 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/605203af-fcdf-42c0-a66f-5c412f8e7770-build-blob-cache\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.575662 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/605203af-fcdf-42c0-a66f-5c412f8e7770-builder-dockercfg-8h4xs-pull\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.575698 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxrc8\" (UniqueName: \"kubernetes.io/projected/605203af-fcdf-42c0-a66f-5c412f8e7770-kube-api-access-gxrc8\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.676862 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/605203af-fcdf-42c0-a66f-5c412f8e7770-builder-dockercfg-8h4xs-pull\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.677191 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxrc8\" (UniqueName: \"kubernetes.io/projected/605203af-fcdf-42c0-a66f-5c412f8e7770-kube-api-access-gxrc8\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.677288 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/605203af-fcdf-42c0-a66f-5c412f8e7770-container-storage-root\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.677395 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/605203af-fcdf-42c0-a66f-5c412f8e7770-node-pullsecrets\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.677479 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/605203af-fcdf-42c0-a66f-5c412f8e7770-buildcachedir\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.677585 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/605203af-fcdf-42c0-a66f-5c412f8e7770-build-system-configs\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.677644 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/605203af-fcdf-42c0-a66f-5c412f8e7770-buildcachedir\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.677669 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/605203af-fcdf-42c0-a66f-5c412f8e7770-buildworkdir\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.677517 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/605203af-fcdf-42c0-a66f-5c412f8e7770-node-pullsecrets\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.677852 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/605203af-fcdf-42c0-a66f-5c412f8e7770-container-storage-root\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.677821 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/605203af-fcdf-42c0-a66f-5c412f8e7770-build-ca-bundles\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.678004 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/605203af-fcdf-42c0-a66f-5c412f8e7770-builder-dockercfg-8h4xs-push\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.678074 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/605203af-fcdf-42c0-a66f-5c412f8e7770-container-storage-run\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.678113 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/605203af-fcdf-42c0-a66f-5c412f8e7770-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.678164 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/605203af-fcdf-42c0-a66f-5c412f8e7770-build-blob-cache\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.678222 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/605203af-fcdf-42c0-a66f-5c412f8e7770-build-system-configs\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.678381 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/605203af-fcdf-42c0-a66f-5c412f8e7770-buildworkdir\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.678983 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/605203af-fcdf-42c0-a66f-5c412f8e7770-container-storage-run\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.679269 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/605203af-fcdf-42c0-a66f-5c412f8e7770-build-blob-cache\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.679330 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/605203af-fcdf-42c0-a66f-5c412f8e7770-build-ca-bundles\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.679907 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/605203af-fcdf-42c0-a66f-5c412f8e7770-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.681940 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/605203af-fcdf-42c0-a66f-5c412f8e7770-builder-dockercfg-8h4xs-pull\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.682807 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/605203af-fcdf-42c0-a66f-5c412f8e7770-builder-dockercfg-8h4xs-push\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.696226 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxrc8\" (UniqueName: \"kubernetes.io/projected/605203af-fcdf-42c0-a66f-5c412f8e7770-kube-api-access-gxrc8\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.704219 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 26 17:22:51 crc kubenswrapper[4856]: I0126 17:22:51.922053 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-2-build"] Jan 26 17:22:52 crc kubenswrapper[4856]: I0126 17:22:52.092208 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"605203af-fcdf-42c0-a66f-5c412f8e7770","Type":"ContainerStarted","Data":"f9ad774bd179b8409f4822d0009683c738837e6f9f7337a5a04fab814edc853d"} Jan 26 17:22:53 crc kubenswrapper[4856]: I0126 17:22:53.100517 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"605203af-fcdf-42c0-a66f-5c412f8e7770","Type":"ContainerStarted","Data":"ddf87c33308074ead375fc61c59fd29203d99ff38dd4ed590fdf9b02056267ff"} Jan 26 17:22:54 crc kubenswrapper[4856]: I0126 17:22:54.109772 4856 generic.go:334] "Generic (PLEG): container finished" podID="605203af-fcdf-42c0-a66f-5c412f8e7770" containerID="ddf87c33308074ead375fc61c59fd29203d99ff38dd4ed590fdf9b02056267ff" exitCode=0 Jan 26 17:22:54 crc kubenswrapper[4856]: I0126 17:22:54.110003 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"605203af-fcdf-42c0-a66f-5c412f8e7770","Type":"ContainerDied","Data":"ddf87c33308074ead375fc61c59fd29203d99ff38dd4ed590fdf9b02056267ff"} Jan 26 17:22:55 crc kubenswrapper[4856]: I0126 17:22:55.117514 4856 generic.go:334] "Generic (PLEG): container finished" podID="605203af-fcdf-42c0-a66f-5c412f8e7770" containerID="5a26cc5ef603c95f249fb40b3e1e055a5dd3ae0f33c81b0f8dfd906d88554058" exitCode=0 Jan 26 17:22:55 crc kubenswrapper[4856]: I0126 17:22:55.117626 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"605203af-fcdf-42c0-a66f-5c412f8e7770","Type":"ContainerDied","Data":"5a26cc5ef603c95f249fb40b3e1e055a5dd3ae0f33c81b0f8dfd906d88554058"} Jan 26 17:22:55 crc kubenswrapper[4856]: I0126 17:22:55.154182 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-webhook-snmp-2-build_605203af-fcdf-42c0-a66f-5c412f8e7770/manage-dockerfile/0.log" Jan 26 17:22:56 crc kubenswrapper[4856]: I0126 17:22:56.127378 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"605203af-fcdf-42c0-a66f-5c412f8e7770","Type":"ContainerStarted","Data":"173ac321b06d4b28b5b12828ba97f3f373de2fefc48808037e0f118d81499c95"} Jan 26 17:22:56 crc kubenswrapper[4856]: I0126 17:22:56.157580 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/prometheus-webhook-snmp-2-build" podStartSLOduration=5.157553043 podStartE2EDuration="5.157553043s" podCreationTimestamp="2026-01-26 17:22:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:22:56.153131486 +0000 UTC m=+1472.106385557" watchObservedRunningTime="2026-01-26 17:22:56.157553043 +0000 UTC m=+1472.110807034" Jan 26 17:22:56 crc kubenswrapper[4856]: I0126 17:22:56.938941 4856 patch_prober.go:28] interesting pod/machine-config-daemon-xm9cq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:22:56 crc kubenswrapper[4856]: I0126 17:22:56.939009 4856 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:23:26 crc kubenswrapper[4856]: I0126 17:23:26.365978 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-webhook-snmp-2-build_605203af-fcdf-42c0-a66f-5c412f8e7770/docker-build/0.log" Jan 26 17:23:26 crc kubenswrapper[4856]: I0126 17:23:26.368390 4856 generic.go:334] "Generic (PLEG): container finished" podID="605203af-fcdf-42c0-a66f-5c412f8e7770" containerID="173ac321b06d4b28b5b12828ba97f3f373de2fefc48808037e0f118d81499c95" exitCode=1 Jan 26 17:23:26 crc kubenswrapper[4856]: I0126 17:23:26.368437 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"605203af-fcdf-42c0-a66f-5c412f8e7770","Type":"ContainerDied","Data":"173ac321b06d4b28b5b12828ba97f3f373de2fefc48808037e0f118d81499c95"} Jan 26 17:23:26 crc kubenswrapper[4856]: I0126 17:23:26.938899 4856 patch_prober.go:28] interesting pod/machine-config-daemon-xm9cq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:23:26 crc kubenswrapper[4856]: I0126 17:23:26.939187 4856 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:23:26 crc kubenswrapper[4856]: I0126 17:23:26.939232 4856 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" Jan 26 17:23:26 crc kubenswrapper[4856]: I0126 17:23:26.939846 4856 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cda3cdbac0b1e3c460ee9a5617b9c5fd59d4db5c67a69b81c9224934be12563c"} pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 26 17:23:26 crc kubenswrapper[4856]: I0126 17:23:26.939902 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" containerName="machine-config-daemon" containerID="cri-o://cda3cdbac0b1e3c460ee9a5617b9c5fd59d4db5c67a69b81c9224934be12563c" gracePeriod=600 Jan 26 17:23:27 crc kubenswrapper[4856]: I0126 17:23:27.379116 4856 generic.go:334] "Generic (PLEG): container finished" podID="63c75ede-5170-4db0-811b-5217ef8d72b3" containerID="cda3cdbac0b1e3c460ee9a5617b9c5fd59d4db5c67a69b81c9224934be12563c" exitCode=0 Jan 26 17:23:27 crc kubenswrapper[4856]: I0126 17:23:27.379311 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" event={"ID":"63c75ede-5170-4db0-811b-5217ef8d72b3","Type":"ContainerDied","Data":"cda3cdbac0b1e3c460ee9a5617b9c5fd59d4db5c67a69b81c9224934be12563c"} Jan 26 17:23:27 crc kubenswrapper[4856]: I0126 17:23:27.380393 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" event={"ID":"63c75ede-5170-4db0-811b-5217ef8d72b3","Type":"ContainerStarted","Data":"b8175a0e79754a858867d9a98f2aa6c52214536db6005b6724cf907eb7a891ee"} Jan 26 17:23:27 crc kubenswrapper[4856]: I0126 17:23:27.380450 4856 scope.go:117] "RemoveContainer" containerID="5846ab4d870be5fcbab796c3e27690d2c13d129480d6fcd21b3b0d1c535f0cff" Jan 26 17:23:27 crc kubenswrapper[4856]: I0126 17:23:27.702859 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-webhook-snmp-2-build_605203af-fcdf-42c0-a66f-5c412f8e7770/docker-build/0.log" Jan 26 17:23:27 crc kubenswrapper[4856]: I0126 17:23:27.704476 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 26 17:23:27 crc kubenswrapper[4856]: I0126 17:23:27.794311 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxrc8\" (UniqueName: \"kubernetes.io/projected/605203af-fcdf-42c0-a66f-5c412f8e7770-kube-api-access-gxrc8\") pod \"605203af-fcdf-42c0-a66f-5c412f8e7770\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " Jan 26 17:23:27 crc kubenswrapper[4856]: I0126 17:23:27.794368 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/605203af-fcdf-42c0-a66f-5c412f8e7770-container-storage-run\") pod \"605203af-fcdf-42c0-a66f-5c412f8e7770\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " Jan 26 17:23:27 crc kubenswrapper[4856]: I0126 17:23:27.794401 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/605203af-fcdf-42c0-a66f-5c412f8e7770-build-system-configs\") pod \"605203af-fcdf-42c0-a66f-5c412f8e7770\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " Jan 26 17:23:27 crc kubenswrapper[4856]: I0126 17:23:27.794445 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/605203af-fcdf-42c0-a66f-5c412f8e7770-node-pullsecrets\") pod \"605203af-fcdf-42c0-a66f-5c412f8e7770\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " Jan 26 17:23:27 crc kubenswrapper[4856]: I0126 17:23:27.794484 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/605203af-fcdf-42c0-a66f-5c412f8e7770-build-proxy-ca-bundles\") pod \"605203af-fcdf-42c0-a66f-5c412f8e7770\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " Jan 26 17:23:27 crc kubenswrapper[4856]: I0126 17:23:27.794547 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/605203af-fcdf-42c0-a66f-5c412f8e7770-build-blob-cache\") pod \"605203af-fcdf-42c0-a66f-5c412f8e7770\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " Jan 26 17:23:27 crc kubenswrapper[4856]: I0126 17:23:27.794593 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/605203af-fcdf-42c0-a66f-5c412f8e7770-builder-dockercfg-8h4xs-pull\") pod \"605203af-fcdf-42c0-a66f-5c412f8e7770\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " Jan 26 17:23:27 crc kubenswrapper[4856]: I0126 17:23:27.794635 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/605203af-fcdf-42c0-a66f-5c412f8e7770-build-ca-bundles\") pod \"605203af-fcdf-42c0-a66f-5c412f8e7770\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " Jan 26 17:23:27 crc kubenswrapper[4856]: I0126 17:23:27.794655 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/605203af-fcdf-42c0-a66f-5c412f8e7770-builder-dockercfg-8h4xs-push\") pod \"605203af-fcdf-42c0-a66f-5c412f8e7770\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " Jan 26 17:23:27 crc kubenswrapper[4856]: I0126 17:23:27.794674 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/605203af-fcdf-42c0-a66f-5c412f8e7770-container-storage-root\") pod \"605203af-fcdf-42c0-a66f-5c412f8e7770\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " Jan 26 17:23:27 crc kubenswrapper[4856]: I0126 17:23:27.794687 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/605203af-fcdf-42c0-a66f-5c412f8e7770-buildcachedir\") pod \"605203af-fcdf-42c0-a66f-5c412f8e7770\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " Jan 26 17:23:27 crc kubenswrapper[4856]: I0126 17:23:27.794721 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/605203af-fcdf-42c0-a66f-5c412f8e7770-buildworkdir\") pod \"605203af-fcdf-42c0-a66f-5c412f8e7770\" (UID: \"605203af-fcdf-42c0-a66f-5c412f8e7770\") " Jan 26 17:23:27 crc kubenswrapper[4856]: I0126 17:23:27.795037 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/605203af-fcdf-42c0-a66f-5c412f8e7770-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "605203af-fcdf-42c0-a66f-5c412f8e7770" (UID: "605203af-fcdf-42c0-a66f-5c412f8e7770"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:23:27 crc kubenswrapper[4856]: I0126 17:23:27.795270 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/605203af-fcdf-42c0-a66f-5c412f8e7770-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "605203af-fcdf-42c0-a66f-5c412f8e7770" (UID: "605203af-fcdf-42c0-a66f-5c412f8e7770"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:23:27 crc kubenswrapper[4856]: I0126 17:23:27.795476 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/605203af-fcdf-42c0-a66f-5c412f8e7770-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "605203af-fcdf-42c0-a66f-5c412f8e7770" (UID: "605203af-fcdf-42c0-a66f-5c412f8e7770"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:23:27 crc kubenswrapper[4856]: I0126 17:23:27.795620 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/605203af-fcdf-42c0-a66f-5c412f8e7770-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "605203af-fcdf-42c0-a66f-5c412f8e7770" (UID: "605203af-fcdf-42c0-a66f-5c412f8e7770"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:23:27 crc kubenswrapper[4856]: I0126 17:23:27.794945 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/605203af-fcdf-42c0-a66f-5c412f8e7770-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "605203af-fcdf-42c0-a66f-5c412f8e7770" (UID: "605203af-fcdf-42c0-a66f-5c412f8e7770"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:23:27 crc kubenswrapper[4856]: I0126 17:23:27.797189 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/605203af-fcdf-42c0-a66f-5c412f8e7770-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "605203af-fcdf-42c0-a66f-5c412f8e7770" (UID: "605203af-fcdf-42c0-a66f-5c412f8e7770"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:23:27 crc kubenswrapper[4856]: I0126 17:23:27.798005 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/605203af-fcdf-42c0-a66f-5c412f8e7770-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "605203af-fcdf-42c0-a66f-5c412f8e7770" (UID: "605203af-fcdf-42c0-a66f-5c412f8e7770"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:23:27 crc kubenswrapper[4856]: I0126 17:23:27.801883 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/605203af-fcdf-42c0-a66f-5c412f8e7770-builder-dockercfg-8h4xs-pull" (OuterVolumeSpecName: "builder-dockercfg-8h4xs-pull") pod "605203af-fcdf-42c0-a66f-5c412f8e7770" (UID: "605203af-fcdf-42c0-a66f-5c412f8e7770"). InnerVolumeSpecName "builder-dockercfg-8h4xs-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:23:27 crc kubenswrapper[4856]: I0126 17:23:27.804148 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/605203af-fcdf-42c0-a66f-5c412f8e7770-kube-api-access-gxrc8" (OuterVolumeSpecName: "kube-api-access-gxrc8") pod "605203af-fcdf-42c0-a66f-5c412f8e7770" (UID: "605203af-fcdf-42c0-a66f-5c412f8e7770"). InnerVolumeSpecName "kube-api-access-gxrc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:23:27 crc kubenswrapper[4856]: I0126 17:23:27.807901 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/605203af-fcdf-42c0-a66f-5c412f8e7770-builder-dockercfg-8h4xs-push" (OuterVolumeSpecName: "builder-dockercfg-8h4xs-push") pod "605203af-fcdf-42c0-a66f-5c412f8e7770" (UID: "605203af-fcdf-42c0-a66f-5c412f8e7770"). InnerVolumeSpecName "builder-dockercfg-8h4xs-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:23:27 crc kubenswrapper[4856]: I0126 17:23:27.896661 4856 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/605203af-fcdf-42c0-a66f-5c412f8e7770-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Jan 26 17:23:27 crc kubenswrapper[4856]: I0126 17:23:27.896914 4856 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/605203af-fcdf-42c0-a66f-5c412f8e7770-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 26 17:23:27 crc kubenswrapper[4856]: I0126 17:23:27.897026 4856 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/605203af-fcdf-42c0-a66f-5c412f8e7770-builder-dockercfg-8h4xs-pull\") on node \"crc\" DevicePath \"\"" Jan 26 17:23:27 crc kubenswrapper[4856]: I0126 17:23:27.897113 4856 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/605203af-fcdf-42c0-a66f-5c412f8e7770-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 26 17:23:27 crc kubenswrapper[4856]: I0126 17:23:27.897193 4856 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/605203af-fcdf-42c0-a66f-5c412f8e7770-builder-dockercfg-8h4xs-push\") on node \"crc\" DevicePath \"\"" Jan 26 17:23:27 crc kubenswrapper[4856]: I0126 17:23:27.897286 4856 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/605203af-fcdf-42c0-a66f-5c412f8e7770-buildcachedir\") on node \"crc\" DevicePath \"\"" Jan 26 17:23:27 crc kubenswrapper[4856]: I0126 17:23:27.897367 4856 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/605203af-fcdf-42c0-a66f-5c412f8e7770-buildworkdir\") on node \"crc\" DevicePath \"\"" Jan 26 17:23:27 crc kubenswrapper[4856]: I0126 17:23:27.897452 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxrc8\" (UniqueName: \"kubernetes.io/projected/605203af-fcdf-42c0-a66f-5c412f8e7770-kube-api-access-gxrc8\") on node \"crc\" DevicePath \"\"" Jan 26 17:23:27 crc kubenswrapper[4856]: I0126 17:23:27.897554 4856 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/605203af-fcdf-42c0-a66f-5c412f8e7770-container-storage-run\") on node \"crc\" DevicePath \"\"" Jan 26 17:23:27 crc kubenswrapper[4856]: I0126 17:23:27.897649 4856 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/605203af-fcdf-42c0-a66f-5c412f8e7770-build-system-configs\") on node \"crc\" DevicePath \"\"" Jan 26 17:23:27 crc kubenswrapper[4856]: I0126 17:23:27.904829 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/605203af-fcdf-42c0-a66f-5c412f8e7770-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "605203af-fcdf-42c0-a66f-5c412f8e7770" (UID: "605203af-fcdf-42c0-a66f-5c412f8e7770"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:23:27 crc kubenswrapper[4856]: I0126 17:23:27.999748 4856 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/605203af-fcdf-42c0-a66f-5c412f8e7770-build-blob-cache\") on node \"crc\" DevicePath \"\"" Jan 26 17:23:28 crc kubenswrapper[4856]: I0126 17:23:28.359959 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/605203af-fcdf-42c0-a66f-5c412f8e7770-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "605203af-fcdf-42c0-a66f-5c412f8e7770" (UID: "605203af-fcdf-42c0-a66f-5c412f8e7770"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:23:28 crc kubenswrapper[4856]: I0126 17:23:28.396602 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-webhook-snmp-2-build_605203af-fcdf-42c0-a66f-5c412f8e7770/docker-build/0.log" Jan 26 17:23:28 crc kubenswrapper[4856]: I0126 17:23:28.398309 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"605203af-fcdf-42c0-a66f-5c412f8e7770","Type":"ContainerDied","Data":"f9ad774bd179b8409f4822d0009683c738837e6f9f7337a5a04fab814edc853d"} Jan 26 17:23:28 crc kubenswrapper[4856]: I0126 17:23:28.398358 4856 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f9ad774bd179b8409f4822d0009683c738837e6f9f7337a5a04fab814edc853d" Jan 26 17:23:28 crc kubenswrapper[4856]: I0126 17:23:28.398446 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-2-build" Jan 26 17:23:28 crc kubenswrapper[4856]: I0126 17:23:28.405463 4856 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/605203af-fcdf-42c0-a66f-5c412f8e7770-container-storage-root\") on node \"crc\" DevicePath \"\"" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.084616 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/prometheus-webhook-snmp-3-build"] Jan 26 17:23:38 crc kubenswrapper[4856]: E0126 17:23:38.086425 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="605203af-fcdf-42c0-a66f-5c412f8e7770" containerName="git-clone" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.086458 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="605203af-fcdf-42c0-a66f-5c412f8e7770" containerName="git-clone" Jan 26 17:23:38 crc kubenswrapper[4856]: E0126 17:23:38.086490 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="605203af-fcdf-42c0-a66f-5c412f8e7770" containerName="docker-build" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.086507 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="605203af-fcdf-42c0-a66f-5c412f8e7770" containerName="docker-build" Jan 26 17:23:38 crc kubenswrapper[4856]: E0126 17:23:38.086564 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="605203af-fcdf-42c0-a66f-5c412f8e7770" containerName="manage-dockerfile" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.086580 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="605203af-fcdf-42c0-a66f-5c412f8e7770" containerName="manage-dockerfile" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.086822 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="605203af-fcdf-42c0-a66f-5c412f8e7770" containerName="docker-build" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.088269 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-3-build" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.090731 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-3-sys-config" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.091972 4856 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-8h4xs" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.093309 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-3-global-ca" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.093775 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-3-ca" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.105618 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-3-build"] Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.269290 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/c15c4956-0479-4646-86c9-ca4a7ce31a28-build-blob-cache\") pod \"prometheus-webhook-snmp-3-build\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " pod="service-telemetry/prometheus-webhook-snmp-3-build" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.269386 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/c15c4956-0479-4646-86c9-ca4a7ce31a28-buildcachedir\") pod \"prometheus-webhook-snmp-3-build\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " pod="service-telemetry/prometheus-webhook-snmp-3-build" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.269455 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/c15c4956-0479-4646-86c9-ca4a7ce31a28-container-storage-run\") pod \"prometheus-webhook-snmp-3-build\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " pod="service-telemetry/prometheus-webhook-snmp-3-build" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.269589 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s28tl\" (UniqueName: \"kubernetes.io/projected/c15c4956-0479-4646-86c9-ca4a7ce31a28-kube-api-access-s28tl\") pod \"prometheus-webhook-snmp-3-build\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " pod="service-telemetry/prometheus-webhook-snmp-3-build" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.269663 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c15c4956-0479-4646-86c9-ca4a7ce31a28-node-pullsecrets\") pod \"prometheus-webhook-snmp-3-build\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " pod="service-telemetry/prometheus-webhook-snmp-3-build" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.269737 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c15c4956-0479-4646-86c9-ca4a7ce31a28-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-3-build\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " pod="service-telemetry/prometheus-webhook-snmp-3-build" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.269825 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c15c4956-0479-4646-86c9-ca4a7ce31a28-build-ca-bundles\") pod \"prometheus-webhook-snmp-3-build\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " pod="service-telemetry/prometheus-webhook-snmp-3-build" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.269894 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/c15c4956-0479-4646-86c9-ca4a7ce31a28-container-storage-root\") pod \"prometheus-webhook-snmp-3-build\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " pod="service-telemetry/prometheus-webhook-snmp-3-build" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.269970 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/c15c4956-0479-4646-86c9-ca4a7ce31a28-builder-dockercfg-8h4xs-pull\") pod \"prometheus-webhook-snmp-3-build\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " pod="service-telemetry/prometheus-webhook-snmp-3-build" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.270029 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/c15c4956-0479-4646-86c9-ca4a7ce31a28-build-system-configs\") pod \"prometheus-webhook-snmp-3-build\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " pod="service-telemetry/prometheus-webhook-snmp-3-build" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.270124 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/c15c4956-0479-4646-86c9-ca4a7ce31a28-builder-dockercfg-8h4xs-push\") pod \"prometheus-webhook-snmp-3-build\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " pod="service-telemetry/prometheus-webhook-snmp-3-build" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.270238 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/c15c4956-0479-4646-86c9-ca4a7ce31a28-buildworkdir\") pod \"prometheus-webhook-snmp-3-build\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " pod="service-telemetry/prometheus-webhook-snmp-3-build" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.371821 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/c15c4956-0479-4646-86c9-ca4a7ce31a28-container-storage-root\") pod \"prometheus-webhook-snmp-3-build\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " pod="service-telemetry/prometheus-webhook-snmp-3-build" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.371880 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/c15c4956-0479-4646-86c9-ca4a7ce31a28-builder-dockercfg-8h4xs-pull\") pod \"prometheus-webhook-snmp-3-build\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " pod="service-telemetry/prometheus-webhook-snmp-3-build" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.371900 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/c15c4956-0479-4646-86c9-ca4a7ce31a28-build-system-configs\") pod \"prometheus-webhook-snmp-3-build\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " pod="service-telemetry/prometheus-webhook-snmp-3-build" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.371922 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/c15c4956-0479-4646-86c9-ca4a7ce31a28-builder-dockercfg-8h4xs-push\") pod \"prometheus-webhook-snmp-3-build\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " pod="service-telemetry/prometheus-webhook-snmp-3-build" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.371961 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/c15c4956-0479-4646-86c9-ca4a7ce31a28-buildworkdir\") pod \"prometheus-webhook-snmp-3-build\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " pod="service-telemetry/prometheus-webhook-snmp-3-build" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.371985 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/c15c4956-0479-4646-86c9-ca4a7ce31a28-build-blob-cache\") pod \"prometheus-webhook-snmp-3-build\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " pod="service-telemetry/prometheus-webhook-snmp-3-build" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.372004 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/c15c4956-0479-4646-86c9-ca4a7ce31a28-buildcachedir\") pod \"prometheus-webhook-snmp-3-build\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " pod="service-telemetry/prometheus-webhook-snmp-3-build" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.372027 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/c15c4956-0479-4646-86c9-ca4a7ce31a28-container-storage-run\") pod \"prometheus-webhook-snmp-3-build\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " pod="service-telemetry/prometheus-webhook-snmp-3-build" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.372042 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s28tl\" (UniqueName: \"kubernetes.io/projected/c15c4956-0479-4646-86c9-ca4a7ce31a28-kube-api-access-s28tl\") pod \"prometheus-webhook-snmp-3-build\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " pod="service-telemetry/prometheus-webhook-snmp-3-build" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.372064 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c15c4956-0479-4646-86c9-ca4a7ce31a28-node-pullsecrets\") pod \"prometheus-webhook-snmp-3-build\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " pod="service-telemetry/prometheus-webhook-snmp-3-build" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.372084 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c15c4956-0479-4646-86c9-ca4a7ce31a28-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-3-build\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " pod="service-telemetry/prometheus-webhook-snmp-3-build" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.372102 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c15c4956-0479-4646-86c9-ca4a7ce31a28-build-ca-bundles\") pod \"prometheus-webhook-snmp-3-build\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " pod="service-telemetry/prometheus-webhook-snmp-3-build" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.372371 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/c15c4956-0479-4646-86c9-ca4a7ce31a28-container-storage-root\") pod \"prometheus-webhook-snmp-3-build\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " pod="service-telemetry/prometheus-webhook-snmp-3-build" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.372433 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/c15c4956-0479-4646-86c9-ca4a7ce31a28-buildcachedir\") pod \"prometheus-webhook-snmp-3-build\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " pod="service-telemetry/prometheus-webhook-snmp-3-build" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.372830 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c15c4956-0479-4646-86c9-ca4a7ce31a28-build-ca-bundles\") pod \"prometheus-webhook-snmp-3-build\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " pod="service-telemetry/prometheus-webhook-snmp-3-build" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.372894 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/c15c4956-0479-4646-86c9-ca4a7ce31a28-build-system-configs\") pod \"prometheus-webhook-snmp-3-build\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " pod="service-telemetry/prometheus-webhook-snmp-3-build" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.373008 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/c15c4956-0479-4646-86c9-ca4a7ce31a28-container-storage-run\") pod \"prometheus-webhook-snmp-3-build\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " pod="service-telemetry/prometheus-webhook-snmp-3-build" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.373006 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c15c4956-0479-4646-86c9-ca4a7ce31a28-node-pullsecrets\") pod \"prometheus-webhook-snmp-3-build\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " pod="service-telemetry/prometheus-webhook-snmp-3-build" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.373186 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/c15c4956-0479-4646-86c9-ca4a7ce31a28-build-blob-cache\") pod \"prometheus-webhook-snmp-3-build\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " pod="service-telemetry/prometheus-webhook-snmp-3-build" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.373660 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c15c4956-0479-4646-86c9-ca4a7ce31a28-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-3-build\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " pod="service-telemetry/prometheus-webhook-snmp-3-build" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.374233 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/c15c4956-0479-4646-86c9-ca4a7ce31a28-buildworkdir\") pod \"prometheus-webhook-snmp-3-build\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " pod="service-telemetry/prometheus-webhook-snmp-3-build" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.378107 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/c15c4956-0479-4646-86c9-ca4a7ce31a28-builder-dockercfg-8h4xs-push\") pod \"prometheus-webhook-snmp-3-build\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " pod="service-telemetry/prometheus-webhook-snmp-3-build" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.389809 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/c15c4956-0479-4646-86c9-ca4a7ce31a28-builder-dockercfg-8h4xs-pull\") pod \"prometheus-webhook-snmp-3-build\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " pod="service-telemetry/prometheus-webhook-snmp-3-build" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.400332 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s28tl\" (UniqueName: \"kubernetes.io/projected/c15c4956-0479-4646-86c9-ca4a7ce31a28-kube-api-access-s28tl\") pod \"prometheus-webhook-snmp-3-build\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " pod="service-telemetry/prometheus-webhook-snmp-3-build" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.409163 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-3-build" Jan 26 17:23:38 crc kubenswrapper[4856]: I0126 17:23:38.644739 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-3-build"] Jan 26 17:23:39 crc kubenswrapper[4856]: I0126 17:23:39.497345 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-3-build" event={"ID":"c15c4956-0479-4646-86c9-ca4a7ce31a28","Type":"ContainerStarted","Data":"187f5c587ecc9aa2039c8f93aecae55132826926fc216d1bfc5a59d963e568e4"} Jan 26 17:23:39 crc kubenswrapper[4856]: I0126 17:23:39.497726 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-3-build" event={"ID":"c15c4956-0479-4646-86c9-ca4a7ce31a28","Type":"ContainerStarted","Data":"04562fdb1bb25c90ae0571bf68dcbe10145ddf06b7e4caf1d7fb642924f80caf"} Jan 26 17:23:40 crc kubenswrapper[4856]: I0126 17:23:40.505610 4856 generic.go:334] "Generic (PLEG): container finished" podID="c15c4956-0479-4646-86c9-ca4a7ce31a28" containerID="187f5c587ecc9aa2039c8f93aecae55132826926fc216d1bfc5a59d963e568e4" exitCode=0 Jan 26 17:23:40 crc kubenswrapper[4856]: I0126 17:23:40.505739 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-3-build" event={"ID":"c15c4956-0479-4646-86c9-ca4a7ce31a28","Type":"ContainerDied","Data":"187f5c587ecc9aa2039c8f93aecae55132826926fc216d1bfc5a59d963e568e4"} Jan 26 17:23:41 crc kubenswrapper[4856]: I0126 17:23:41.515013 4856 generic.go:334] "Generic (PLEG): container finished" podID="c15c4956-0479-4646-86c9-ca4a7ce31a28" containerID="4f96c5d3e3aa29c56225101e8cebf73a3825ff877c4c07e8e94cd75366ff9736" exitCode=0 Jan 26 17:23:41 crc kubenswrapper[4856]: I0126 17:23:41.515063 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-3-build" event={"ID":"c15c4956-0479-4646-86c9-ca4a7ce31a28","Type":"ContainerDied","Data":"4f96c5d3e3aa29c56225101e8cebf73a3825ff877c4c07e8e94cd75366ff9736"} Jan 26 17:23:41 crc kubenswrapper[4856]: I0126 17:23:41.564218 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-webhook-snmp-3-build_c15c4956-0479-4646-86c9-ca4a7ce31a28/manage-dockerfile/0.log" Jan 26 17:23:42 crc kubenswrapper[4856]: I0126 17:23:42.525313 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-3-build" event={"ID":"c15c4956-0479-4646-86c9-ca4a7ce31a28","Type":"ContainerStarted","Data":"3ecfe1e65b08426eb861eac05b29172d1abeeaf5b3a7ab21586bd8ce360d51d3"} Jan 26 17:23:42 crc kubenswrapper[4856]: I0126 17:23:42.621877 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/prometheus-webhook-snmp-3-build" podStartSLOduration=5.621838745 podStartE2EDuration="5.621838745s" podCreationTimestamp="2026-01-26 17:23:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:23:42.588935834 +0000 UTC m=+1518.542189835" watchObservedRunningTime="2026-01-26 17:23:42.621838745 +0000 UTC m=+1518.575092726" Jan 26 17:23:58 crc kubenswrapper[4856]: I0126 17:23:58.407096 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hpsnz"] Jan 26 17:23:58 crc kubenswrapper[4856]: I0126 17:23:58.409431 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hpsnz" Jan 26 17:23:58 crc kubenswrapper[4856]: I0126 17:23:58.421633 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hpsnz"] Jan 26 17:23:58 crc kubenswrapper[4856]: I0126 17:23:58.476422 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07f01f78-5312-4b86-8c59-2c02f054f99d-utilities\") pod \"redhat-operators-hpsnz\" (UID: \"07f01f78-5312-4b86-8c59-2c02f054f99d\") " pod="openshift-marketplace/redhat-operators-hpsnz" Jan 26 17:23:58 crc kubenswrapper[4856]: I0126 17:23:58.476593 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07f01f78-5312-4b86-8c59-2c02f054f99d-catalog-content\") pod \"redhat-operators-hpsnz\" (UID: \"07f01f78-5312-4b86-8c59-2c02f054f99d\") " pod="openshift-marketplace/redhat-operators-hpsnz" Jan 26 17:23:58 crc kubenswrapper[4856]: I0126 17:23:58.476630 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlj8b\" (UniqueName: \"kubernetes.io/projected/07f01f78-5312-4b86-8c59-2c02f054f99d-kube-api-access-xlj8b\") pod \"redhat-operators-hpsnz\" (UID: \"07f01f78-5312-4b86-8c59-2c02f054f99d\") " pod="openshift-marketplace/redhat-operators-hpsnz" Jan 26 17:23:58 crc kubenswrapper[4856]: I0126 17:23:58.578155 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlj8b\" (UniqueName: \"kubernetes.io/projected/07f01f78-5312-4b86-8c59-2c02f054f99d-kube-api-access-xlj8b\") pod \"redhat-operators-hpsnz\" (UID: \"07f01f78-5312-4b86-8c59-2c02f054f99d\") " pod="openshift-marketplace/redhat-operators-hpsnz" Jan 26 17:23:58 crc kubenswrapper[4856]: I0126 17:23:58.578303 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07f01f78-5312-4b86-8c59-2c02f054f99d-utilities\") pod \"redhat-operators-hpsnz\" (UID: \"07f01f78-5312-4b86-8c59-2c02f054f99d\") " pod="openshift-marketplace/redhat-operators-hpsnz" Jan 26 17:23:58 crc kubenswrapper[4856]: I0126 17:23:58.578376 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07f01f78-5312-4b86-8c59-2c02f054f99d-catalog-content\") pod \"redhat-operators-hpsnz\" (UID: \"07f01f78-5312-4b86-8c59-2c02f054f99d\") " pod="openshift-marketplace/redhat-operators-hpsnz" Jan 26 17:23:58 crc kubenswrapper[4856]: I0126 17:23:58.578932 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07f01f78-5312-4b86-8c59-2c02f054f99d-catalog-content\") pod \"redhat-operators-hpsnz\" (UID: \"07f01f78-5312-4b86-8c59-2c02f054f99d\") " pod="openshift-marketplace/redhat-operators-hpsnz" Jan 26 17:23:58 crc kubenswrapper[4856]: I0126 17:23:58.579079 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07f01f78-5312-4b86-8c59-2c02f054f99d-utilities\") pod \"redhat-operators-hpsnz\" (UID: \"07f01f78-5312-4b86-8c59-2c02f054f99d\") " pod="openshift-marketplace/redhat-operators-hpsnz" Jan 26 17:23:58 crc kubenswrapper[4856]: I0126 17:23:58.597902 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlj8b\" (UniqueName: \"kubernetes.io/projected/07f01f78-5312-4b86-8c59-2c02f054f99d-kube-api-access-xlj8b\") pod \"redhat-operators-hpsnz\" (UID: \"07f01f78-5312-4b86-8c59-2c02f054f99d\") " pod="openshift-marketplace/redhat-operators-hpsnz" Jan 26 17:23:58 crc kubenswrapper[4856]: I0126 17:23:58.730063 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hpsnz" Jan 26 17:23:59 crc kubenswrapper[4856]: I0126 17:23:59.131164 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hpsnz"] Jan 26 17:23:59 crc kubenswrapper[4856]: I0126 17:23:59.656023 4856 generic.go:334] "Generic (PLEG): container finished" podID="07f01f78-5312-4b86-8c59-2c02f054f99d" containerID="65ff28a459f4ead12ab9875ba11141cd0dd1d47926b020bafc7cd061527be0b4" exitCode=0 Jan 26 17:23:59 crc kubenswrapper[4856]: I0126 17:23:59.656069 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hpsnz" event={"ID":"07f01f78-5312-4b86-8c59-2c02f054f99d","Type":"ContainerDied","Data":"65ff28a459f4ead12ab9875ba11141cd0dd1d47926b020bafc7cd061527be0b4"} Jan 26 17:23:59 crc kubenswrapper[4856]: I0126 17:23:59.656097 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hpsnz" event={"ID":"07f01f78-5312-4b86-8c59-2c02f054f99d","Type":"ContainerStarted","Data":"2b7f525a65ab979c2adc1221856dc62b10204477ecabfc28c50af78293c15e81"} Jan 26 17:23:59 crc kubenswrapper[4856]: I0126 17:23:59.658577 4856 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 26 17:24:00 crc kubenswrapper[4856]: I0126 17:24:00.665547 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hpsnz" event={"ID":"07f01f78-5312-4b86-8c59-2c02f054f99d","Type":"ContainerStarted","Data":"4c9c83f945a435941423e915060ee89b3979fe0baafd5195d79d8cf074728770"} Jan 26 17:24:01 crc kubenswrapper[4856]: I0126 17:24:01.677055 4856 generic.go:334] "Generic (PLEG): container finished" podID="07f01f78-5312-4b86-8c59-2c02f054f99d" containerID="4c9c83f945a435941423e915060ee89b3979fe0baafd5195d79d8cf074728770" exitCode=0 Jan 26 17:24:01 crc kubenswrapper[4856]: I0126 17:24:01.677180 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hpsnz" event={"ID":"07f01f78-5312-4b86-8c59-2c02f054f99d","Type":"ContainerDied","Data":"4c9c83f945a435941423e915060ee89b3979fe0baafd5195d79d8cf074728770"} Jan 26 17:24:02 crc kubenswrapper[4856]: I0126 17:24:02.686702 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hpsnz" event={"ID":"07f01f78-5312-4b86-8c59-2c02f054f99d","Type":"ContainerStarted","Data":"13ab3de3d22799488895b01d04fac36f5b9da2ddb5cd8acff648fd53826c28b3"} Jan 26 17:24:03 crc kubenswrapper[4856]: I0126 17:24:03.715227 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hpsnz" podStartSLOduration=3.152016927 podStartE2EDuration="5.715200982s" podCreationTimestamp="2026-01-26 17:23:58 +0000 UTC" firstStartedPulling="2026-01-26 17:23:59.658286876 +0000 UTC m=+1535.611540857" lastFinishedPulling="2026-01-26 17:24:02.221470931 +0000 UTC m=+1538.174724912" observedRunningTime="2026-01-26 17:24:03.713829533 +0000 UTC m=+1539.667083584" watchObservedRunningTime="2026-01-26 17:24:03.715200982 +0000 UTC m=+1539.668454973" Jan 26 17:24:08 crc kubenswrapper[4856]: I0126 17:24:08.731100 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hpsnz" Jan 26 17:24:08 crc kubenswrapper[4856]: I0126 17:24:08.731717 4856 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hpsnz" Jan 26 17:24:08 crc kubenswrapper[4856]: I0126 17:24:08.776252 4856 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hpsnz" Jan 26 17:24:09 crc kubenswrapper[4856]: I0126 17:24:09.777407 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hpsnz" Jan 26 17:24:09 crc kubenswrapper[4856]: I0126 17:24:09.823033 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hpsnz"] Jan 26 17:24:11 crc kubenswrapper[4856]: I0126 17:24:11.745371 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hpsnz" podUID="07f01f78-5312-4b86-8c59-2c02f054f99d" containerName="registry-server" containerID="cri-o://13ab3de3d22799488895b01d04fac36f5b9da2ddb5cd8acff648fd53826c28b3" gracePeriod=2 Jan 26 17:24:15 crc kubenswrapper[4856]: I0126 17:24:15.659942 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hpsnz" Jan 26 17:24:15 crc kubenswrapper[4856]: I0126 17:24:15.726726 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07f01f78-5312-4b86-8c59-2c02f054f99d-utilities\") pod \"07f01f78-5312-4b86-8c59-2c02f054f99d\" (UID: \"07f01f78-5312-4b86-8c59-2c02f054f99d\") " Jan 26 17:24:15 crc kubenswrapper[4856]: I0126 17:24:15.726900 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xlj8b\" (UniqueName: \"kubernetes.io/projected/07f01f78-5312-4b86-8c59-2c02f054f99d-kube-api-access-xlj8b\") pod \"07f01f78-5312-4b86-8c59-2c02f054f99d\" (UID: \"07f01f78-5312-4b86-8c59-2c02f054f99d\") " Jan 26 17:24:15 crc kubenswrapper[4856]: I0126 17:24:15.727003 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07f01f78-5312-4b86-8c59-2c02f054f99d-catalog-content\") pod \"07f01f78-5312-4b86-8c59-2c02f054f99d\" (UID: \"07f01f78-5312-4b86-8c59-2c02f054f99d\") " Jan 26 17:24:15 crc kubenswrapper[4856]: I0126 17:24:15.732067 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07f01f78-5312-4b86-8c59-2c02f054f99d-utilities" (OuterVolumeSpecName: "utilities") pod "07f01f78-5312-4b86-8c59-2c02f054f99d" (UID: "07f01f78-5312-4b86-8c59-2c02f054f99d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:24:15 crc kubenswrapper[4856]: I0126 17:24:15.737416 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07f01f78-5312-4b86-8c59-2c02f054f99d-kube-api-access-xlj8b" (OuterVolumeSpecName: "kube-api-access-xlj8b") pod "07f01f78-5312-4b86-8c59-2c02f054f99d" (UID: "07f01f78-5312-4b86-8c59-2c02f054f99d"). InnerVolumeSpecName "kube-api-access-xlj8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:24:15 crc kubenswrapper[4856]: I0126 17:24:15.786194 4856 generic.go:334] "Generic (PLEG): container finished" podID="07f01f78-5312-4b86-8c59-2c02f054f99d" containerID="13ab3de3d22799488895b01d04fac36f5b9da2ddb5cd8acff648fd53826c28b3" exitCode=0 Jan 26 17:24:15 crc kubenswrapper[4856]: I0126 17:24:15.786280 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hpsnz" event={"ID":"07f01f78-5312-4b86-8c59-2c02f054f99d","Type":"ContainerDied","Data":"13ab3de3d22799488895b01d04fac36f5b9da2ddb5cd8acff648fd53826c28b3"} Jan 26 17:24:15 crc kubenswrapper[4856]: I0126 17:24:15.786401 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hpsnz" event={"ID":"07f01f78-5312-4b86-8c59-2c02f054f99d","Type":"ContainerDied","Data":"2b7f525a65ab979c2adc1221856dc62b10204477ecabfc28c50af78293c15e81"} Jan 26 17:24:15 crc kubenswrapper[4856]: I0126 17:24:15.786384 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hpsnz" Jan 26 17:24:15 crc kubenswrapper[4856]: I0126 17:24:15.786433 4856 scope.go:117] "RemoveContainer" containerID="13ab3de3d22799488895b01d04fac36f5b9da2ddb5cd8acff648fd53826c28b3" Jan 26 17:24:15 crc kubenswrapper[4856]: I0126 17:24:15.814665 4856 scope.go:117] "RemoveContainer" containerID="4c9c83f945a435941423e915060ee89b3979fe0baafd5195d79d8cf074728770" Jan 26 17:24:15 crc kubenswrapper[4856]: I0126 17:24:15.829591 4856 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07f01f78-5312-4b86-8c59-2c02f054f99d-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 17:24:15 crc kubenswrapper[4856]: I0126 17:24:15.829695 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xlj8b\" (UniqueName: \"kubernetes.io/projected/07f01f78-5312-4b86-8c59-2c02f054f99d-kube-api-access-xlj8b\") on node \"crc\" DevicePath \"\"" Jan 26 17:24:15 crc kubenswrapper[4856]: I0126 17:24:15.846137 4856 scope.go:117] "RemoveContainer" containerID="65ff28a459f4ead12ab9875ba11141cd0dd1d47926b020bafc7cd061527be0b4" Jan 26 17:24:15 crc kubenswrapper[4856]: I0126 17:24:15.883091 4856 scope.go:117] "RemoveContainer" containerID="13ab3de3d22799488895b01d04fac36f5b9da2ddb5cd8acff648fd53826c28b3" Jan 26 17:24:15 crc kubenswrapper[4856]: E0126 17:24:15.885023 4856 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13ab3de3d22799488895b01d04fac36f5b9da2ddb5cd8acff648fd53826c28b3\": container with ID starting with 13ab3de3d22799488895b01d04fac36f5b9da2ddb5cd8acff648fd53826c28b3 not found: ID does not exist" containerID="13ab3de3d22799488895b01d04fac36f5b9da2ddb5cd8acff648fd53826c28b3" Jan 26 17:24:15 crc kubenswrapper[4856]: I0126 17:24:15.885089 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13ab3de3d22799488895b01d04fac36f5b9da2ddb5cd8acff648fd53826c28b3"} err="failed to get container status \"13ab3de3d22799488895b01d04fac36f5b9da2ddb5cd8acff648fd53826c28b3\": rpc error: code = NotFound desc = could not find container \"13ab3de3d22799488895b01d04fac36f5b9da2ddb5cd8acff648fd53826c28b3\": container with ID starting with 13ab3de3d22799488895b01d04fac36f5b9da2ddb5cd8acff648fd53826c28b3 not found: ID does not exist" Jan 26 17:24:15 crc kubenswrapper[4856]: I0126 17:24:15.885121 4856 scope.go:117] "RemoveContainer" containerID="4c9c83f945a435941423e915060ee89b3979fe0baafd5195d79d8cf074728770" Jan 26 17:24:15 crc kubenswrapper[4856]: E0126 17:24:15.887037 4856 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c9c83f945a435941423e915060ee89b3979fe0baafd5195d79d8cf074728770\": container with ID starting with 4c9c83f945a435941423e915060ee89b3979fe0baafd5195d79d8cf074728770 not found: ID does not exist" containerID="4c9c83f945a435941423e915060ee89b3979fe0baafd5195d79d8cf074728770" Jan 26 17:24:15 crc kubenswrapper[4856]: I0126 17:24:15.887095 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c9c83f945a435941423e915060ee89b3979fe0baafd5195d79d8cf074728770"} err="failed to get container status \"4c9c83f945a435941423e915060ee89b3979fe0baafd5195d79d8cf074728770\": rpc error: code = NotFound desc = could not find container \"4c9c83f945a435941423e915060ee89b3979fe0baafd5195d79d8cf074728770\": container with ID starting with 4c9c83f945a435941423e915060ee89b3979fe0baafd5195d79d8cf074728770 not found: ID does not exist" Jan 26 17:24:15 crc kubenswrapper[4856]: I0126 17:24:15.887149 4856 scope.go:117] "RemoveContainer" containerID="65ff28a459f4ead12ab9875ba11141cd0dd1d47926b020bafc7cd061527be0b4" Jan 26 17:24:15 crc kubenswrapper[4856]: E0126 17:24:15.888935 4856 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65ff28a459f4ead12ab9875ba11141cd0dd1d47926b020bafc7cd061527be0b4\": container with ID starting with 65ff28a459f4ead12ab9875ba11141cd0dd1d47926b020bafc7cd061527be0b4 not found: ID does not exist" containerID="65ff28a459f4ead12ab9875ba11141cd0dd1d47926b020bafc7cd061527be0b4" Jan 26 17:24:15 crc kubenswrapper[4856]: I0126 17:24:15.889030 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65ff28a459f4ead12ab9875ba11141cd0dd1d47926b020bafc7cd061527be0b4"} err="failed to get container status \"65ff28a459f4ead12ab9875ba11141cd0dd1d47926b020bafc7cd061527be0b4\": rpc error: code = NotFound desc = could not find container \"65ff28a459f4ead12ab9875ba11141cd0dd1d47926b020bafc7cd061527be0b4\": container with ID starting with 65ff28a459f4ead12ab9875ba11141cd0dd1d47926b020bafc7cd061527be0b4 not found: ID does not exist" Jan 26 17:24:15 crc kubenswrapper[4856]: I0126 17:24:15.901793 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07f01f78-5312-4b86-8c59-2c02f054f99d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "07f01f78-5312-4b86-8c59-2c02f054f99d" (UID: "07f01f78-5312-4b86-8c59-2c02f054f99d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:24:15 crc kubenswrapper[4856]: I0126 17:24:15.931990 4856 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07f01f78-5312-4b86-8c59-2c02f054f99d-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 17:24:16 crc kubenswrapper[4856]: I0126 17:24:16.127178 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hpsnz"] Jan 26 17:24:16 crc kubenswrapper[4856]: I0126 17:24:16.135669 4856 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hpsnz"] Jan 26 17:24:16 crc kubenswrapper[4856]: I0126 17:24:16.797372 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-webhook-snmp-3-build_c15c4956-0479-4646-86c9-ca4a7ce31a28/docker-build/0.log" Jan 26 17:24:16 crc kubenswrapper[4856]: I0126 17:24:16.799666 4856 generic.go:334] "Generic (PLEG): container finished" podID="c15c4956-0479-4646-86c9-ca4a7ce31a28" containerID="3ecfe1e65b08426eb861eac05b29172d1abeeaf5b3a7ab21586bd8ce360d51d3" exitCode=1 Jan 26 17:24:16 crc kubenswrapper[4856]: I0126 17:24:16.799799 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-3-build" event={"ID":"c15c4956-0479-4646-86c9-ca4a7ce31a28","Type":"ContainerDied","Data":"3ecfe1e65b08426eb861eac05b29172d1abeeaf5b3a7ab21586bd8ce360d51d3"} Jan 26 17:24:17 crc kubenswrapper[4856]: I0126 17:24:17.406128 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07f01f78-5312-4b86-8c59-2c02f054f99d" path="/var/lib/kubelet/pods/07f01f78-5312-4b86-8c59-2c02f054f99d/volumes" Jan 26 17:24:18 crc kubenswrapper[4856]: I0126 17:24:18.115398 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-webhook-snmp-3-build_c15c4956-0479-4646-86c9-ca4a7ce31a28/docker-build/0.log" Jan 26 17:24:18 crc kubenswrapper[4856]: I0126 17:24:18.116939 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-3-build" Jan 26 17:24:18 crc kubenswrapper[4856]: I0126 17:24:18.168267 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/c15c4956-0479-4646-86c9-ca4a7ce31a28-buildcachedir\") pod \"c15c4956-0479-4646-86c9-ca4a7ce31a28\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " Jan 26 17:24:18 crc kubenswrapper[4856]: I0126 17:24:18.168389 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/c15c4956-0479-4646-86c9-ca4a7ce31a28-buildworkdir\") pod \"c15c4956-0479-4646-86c9-ca4a7ce31a28\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " Jan 26 17:24:18 crc kubenswrapper[4856]: I0126 17:24:18.168426 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/c15c4956-0479-4646-86c9-ca4a7ce31a28-build-blob-cache\") pod \"c15c4956-0479-4646-86c9-ca4a7ce31a28\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " Jan 26 17:24:18 crc kubenswrapper[4856]: I0126 17:24:18.168454 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/c15c4956-0479-4646-86c9-ca4a7ce31a28-builder-dockercfg-8h4xs-push\") pod \"c15c4956-0479-4646-86c9-ca4a7ce31a28\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " Jan 26 17:24:18 crc kubenswrapper[4856]: I0126 17:24:18.168487 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c15c4956-0479-4646-86c9-ca4a7ce31a28-build-proxy-ca-bundles\") pod \"c15c4956-0479-4646-86c9-ca4a7ce31a28\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " Jan 26 17:24:18 crc kubenswrapper[4856]: I0126 17:24:18.168503 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s28tl\" (UniqueName: \"kubernetes.io/projected/c15c4956-0479-4646-86c9-ca4a7ce31a28-kube-api-access-s28tl\") pod \"c15c4956-0479-4646-86c9-ca4a7ce31a28\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " Jan 26 17:24:18 crc kubenswrapper[4856]: I0126 17:24:18.168524 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c15c4956-0479-4646-86c9-ca4a7ce31a28-build-ca-bundles\") pod \"c15c4956-0479-4646-86c9-ca4a7ce31a28\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " Jan 26 17:24:18 crc kubenswrapper[4856]: I0126 17:24:18.168556 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/c15c4956-0479-4646-86c9-ca4a7ce31a28-builder-dockercfg-8h4xs-pull\") pod \"c15c4956-0479-4646-86c9-ca4a7ce31a28\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " Jan 26 17:24:18 crc kubenswrapper[4856]: I0126 17:24:18.168589 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/c15c4956-0479-4646-86c9-ca4a7ce31a28-container-storage-root\") pod \"c15c4956-0479-4646-86c9-ca4a7ce31a28\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " Jan 26 17:24:18 crc kubenswrapper[4856]: I0126 17:24:18.168623 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c15c4956-0479-4646-86c9-ca4a7ce31a28-node-pullsecrets\") pod \"c15c4956-0479-4646-86c9-ca4a7ce31a28\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " Jan 26 17:24:18 crc kubenswrapper[4856]: I0126 17:24:18.168669 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/c15c4956-0479-4646-86c9-ca4a7ce31a28-container-storage-run\") pod \"c15c4956-0479-4646-86c9-ca4a7ce31a28\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " Jan 26 17:24:18 crc kubenswrapper[4856]: I0126 17:24:18.168781 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/c15c4956-0479-4646-86c9-ca4a7ce31a28-build-system-configs\") pod \"c15c4956-0479-4646-86c9-ca4a7ce31a28\" (UID: \"c15c4956-0479-4646-86c9-ca4a7ce31a28\") " Jan 26 17:24:18 crc kubenswrapper[4856]: I0126 17:24:18.168835 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c15c4956-0479-4646-86c9-ca4a7ce31a28-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "c15c4956-0479-4646-86c9-ca4a7ce31a28" (UID: "c15c4956-0479-4646-86c9-ca4a7ce31a28"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:24:18 crc kubenswrapper[4856]: I0126 17:24:18.169903 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c15c4956-0479-4646-86c9-ca4a7ce31a28-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "c15c4956-0479-4646-86c9-ca4a7ce31a28" (UID: "c15c4956-0479-4646-86c9-ca4a7ce31a28"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:24:18 crc kubenswrapper[4856]: I0126 17:24:18.170818 4856 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/c15c4956-0479-4646-86c9-ca4a7ce31a28-buildcachedir\") on node \"crc\" DevicePath \"\"" Jan 26 17:24:18 crc kubenswrapper[4856]: I0126 17:24:18.170847 4856 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c15c4956-0479-4646-86c9-ca4a7ce31a28-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Jan 26 17:24:18 crc kubenswrapper[4856]: I0126 17:24:18.171180 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c15c4956-0479-4646-86c9-ca4a7ce31a28-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "c15c4956-0479-4646-86c9-ca4a7ce31a28" (UID: "c15c4956-0479-4646-86c9-ca4a7ce31a28"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:24:18 crc kubenswrapper[4856]: I0126 17:24:18.171799 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c15c4956-0479-4646-86c9-ca4a7ce31a28-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "c15c4956-0479-4646-86c9-ca4a7ce31a28" (UID: "c15c4956-0479-4646-86c9-ca4a7ce31a28"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:24:18 crc kubenswrapper[4856]: I0126 17:24:18.171802 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c15c4956-0479-4646-86c9-ca4a7ce31a28-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "c15c4956-0479-4646-86c9-ca4a7ce31a28" (UID: "c15c4956-0479-4646-86c9-ca4a7ce31a28"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:24:18 crc kubenswrapper[4856]: I0126 17:24:18.171980 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c15c4956-0479-4646-86c9-ca4a7ce31a28-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "c15c4956-0479-4646-86c9-ca4a7ce31a28" (UID: "c15c4956-0479-4646-86c9-ca4a7ce31a28"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:24:18 crc kubenswrapper[4856]: I0126 17:24:18.176378 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c15c4956-0479-4646-86c9-ca4a7ce31a28-kube-api-access-s28tl" (OuterVolumeSpecName: "kube-api-access-s28tl") pod "c15c4956-0479-4646-86c9-ca4a7ce31a28" (UID: "c15c4956-0479-4646-86c9-ca4a7ce31a28"). InnerVolumeSpecName "kube-api-access-s28tl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:24:18 crc kubenswrapper[4856]: I0126 17:24:18.177113 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c15c4956-0479-4646-86c9-ca4a7ce31a28-builder-dockercfg-8h4xs-push" (OuterVolumeSpecName: "builder-dockercfg-8h4xs-push") pod "c15c4956-0479-4646-86c9-ca4a7ce31a28" (UID: "c15c4956-0479-4646-86c9-ca4a7ce31a28"). InnerVolumeSpecName "builder-dockercfg-8h4xs-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:24:18 crc kubenswrapper[4856]: I0126 17:24:18.177767 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c15c4956-0479-4646-86c9-ca4a7ce31a28-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "c15c4956-0479-4646-86c9-ca4a7ce31a28" (UID: "c15c4956-0479-4646-86c9-ca4a7ce31a28"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:24:18 crc kubenswrapper[4856]: I0126 17:24:18.178719 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c15c4956-0479-4646-86c9-ca4a7ce31a28-builder-dockercfg-8h4xs-pull" (OuterVolumeSpecName: "builder-dockercfg-8h4xs-pull") pod "c15c4956-0479-4646-86c9-ca4a7ce31a28" (UID: "c15c4956-0479-4646-86c9-ca4a7ce31a28"). InnerVolumeSpecName "builder-dockercfg-8h4xs-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:24:18 crc kubenswrapper[4856]: I0126 17:24:18.259727 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c15c4956-0479-4646-86c9-ca4a7ce31a28-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "c15c4956-0479-4646-86c9-ca4a7ce31a28" (UID: "c15c4956-0479-4646-86c9-ca4a7ce31a28"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:24:18 crc kubenswrapper[4856]: I0126 17:24:18.272256 4856 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/c15c4956-0479-4646-86c9-ca4a7ce31a28-container-storage-run\") on node \"crc\" DevicePath \"\"" Jan 26 17:24:18 crc kubenswrapper[4856]: I0126 17:24:18.272292 4856 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/c15c4956-0479-4646-86c9-ca4a7ce31a28-build-system-configs\") on node \"crc\" DevicePath \"\"" Jan 26 17:24:18 crc kubenswrapper[4856]: I0126 17:24:18.272302 4856 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/c15c4956-0479-4646-86c9-ca4a7ce31a28-buildworkdir\") on node \"crc\" DevicePath \"\"" Jan 26 17:24:18 crc kubenswrapper[4856]: I0126 17:24:18.272312 4856 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/c15c4956-0479-4646-86c9-ca4a7ce31a28-build-blob-cache\") on node \"crc\" DevicePath \"\"" Jan 26 17:24:18 crc kubenswrapper[4856]: I0126 17:24:18.272321 4856 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/c15c4956-0479-4646-86c9-ca4a7ce31a28-builder-dockercfg-8h4xs-push\") on node \"crc\" DevicePath \"\"" Jan 26 17:24:18 crc kubenswrapper[4856]: I0126 17:24:18.272330 4856 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c15c4956-0479-4646-86c9-ca4a7ce31a28-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 26 17:24:18 crc kubenswrapper[4856]: I0126 17:24:18.272338 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s28tl\" (UniqueName: \"kubernetes.io/projected/c15c4956-0479-4646-86c9-ca4a7ce31a28-kube-api-access-s28tl\") on node \"crc\" DevicePath \"\"" Jan 26 17:24:18 crc kubenswrapper[4856]: I0126 17:24:18.272347 4856 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/c15c4956-0479-4646-86c9-ca4a7ce31a28-builder-dockercfg-8h4xs-pull\") on node \"crc\" DevicePath \"\"" Jan 26 17:24:18 crc kubenswrapper[4856]: I0126 17:24:18.272356 4856 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c15c4956-0479-4646-86c9-ca4a7ce31a28-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 26 17:24:18 crc kubenswrapper[4856]: I0126 17:24:18.827673 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-webhook-snmp-3-build_c15c4956-0479-4646-86c9-ca4a7ce31a28/docker-build/0.log" Jan 26 17:24:18 crc kubenswrapper[4856]: I0126 17:24:18.830550 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-3-build" event={"ID":"c15c4956-0479-4646-86c9-ca4a7ce31a28","Type":"ContainerDied","Data":"04562fdb1bb25c90ae0571bf68dcbe10145ddf06b7e4caf1d7fb642924f80caf"} Jan 26 17:24:18 crc kubenswrapper[4856]: I0126 17:24:18.830615 4856 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04562fdb1bb25c90ae0571bf68dcbe10145ddf06b7e4caf1d7fb642924f80caf" Jan 26 17:24:18 crc kubenswrapper[4856]: I0126 17:24:18.830751 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-3-build" Jan 26 17:24:18 crc kubenswrapper[4856]: I0126 17:24:18.853794 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c15c4956-0479-4646-86c9-ca4a7ce31a28-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "c15c4956-0479-4646-86c9-ca4a7ce31a28" (UID: "c15c4956-0479-4646-86c9-ca4a7ce31a28"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:24:18 crc kubenswrapper[4856]: I0126 17:24:18.890294 4856 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/c15c4956-0479-4646-86c9-ca4a7ce31a28-container-storage-root\") on node \"crc\" DevicePath \"\"" Jan 26 17:24:28 crc kubenswrapper[4856]: I0126 17:24:28.143868 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8ttkg"] Jan 26 17:24:28 crc kubenswrapper[4856]: E0126 17:24:28.144837 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c15c4956-0479-4646-86c9-ca4a7ce31a28" containerName="docker-build" Jan 26 17:24:28 crc kubenswrapper[4856]: I0126 17:24:28.144857 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="c15c4956-0479-4646-86c9-ca4a7ce31a28" containerName="docker-build" Jan 26 17:24:28 crc kubenswrapper[4856]: E0126 17:24:28.144878 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07f01f78-5312-4b86-8c59-2c02f054f99d" containerName="registry-server" Jan 26 17:24:28 crc kubenswrapper[4856]: I0126 17:24:28.144888 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="07f01f78-5312-4b86-8c59-2c02f054f99d" containerName="registry-server" Jan 26 17:24:28 crc kubenswrapper[4856]: E0126 17:24:28.144906 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c15c4956-0479-4646-86c9-ca4a7ce31a28" containerName="manage-dockerfile" Jan 26 17:24:28 crc kubenswrapper[4856]: I0126 17:24:28.144914 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="c15c4956-0479-4646-86c9-ca4a7ce31a28" containerName="manage-dockerfile" Jan 26 17:24:28 crc kubenswrapper[4856]: E0126 17:24:28.144930 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07f01f78-5312-4b86-8c59-2c02f054f99d" containerName="extract-utilities" Jan 26 17:24:28 crc kubenswrapper[4856]: I0126 17:24:28.144937 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="07f01f78-5312-4b86-8c59-2c02f054f99d" containerName="extract-utilities" Jan 26 17:24:28 crc kubenswrapper[4856]: E0126 17:24:28.144950 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c15c4956-0479-4646-86c9-ca4a7ce31a28" containerName="git-clone" Jan 26 17:24:28 crc kubenswrapper[4856]: I0126 17:24:28.144960 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="c15c4956-0479-4646-86c9-ca4a7ce31a28" containerName="git-clone" Jan 26 17:24:28 crc kubenswrapper[4856]: E0126 17:24:28.144985 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07f01f78-5312-4b86-8c59-2c02f054f99d" containerName="extract-content" Jan 26 17:24:28 crc kubenswrapper[4856]: I0126 17:24:28.144995 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="07f01f78-5312-4b86-8c59-2c02f054f99d" containerName="extract-content" Jan 26 17:24:28 crc kubenswrapper[4856]: I0126 17:24:28.145127 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="07f01f78-5312-4b86-8c59-2c02f054f99d" containerName="registry-server" Jan 26 17:24:28 crc kubenswrapper[4856]: I0126 17:24:28.145141 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="c15c4956-0479-4646-86c9-ca4a7ce31a28" containerName="docker-build" Jan 26 17:24:28 crc kubenswrapper[4856]: I0126 17:24:28.146220 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8ttkg" Jan 26 17:24:28 crc kubenswrapper[4856]: I0126 17:24:28.156298 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8ttkg"] Jan 26 17:24:28 crc kubenswrapper[4856]: I0126 17:24:28.236769 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bddzx\" (UniqueName: \"kubernetes.io/projected/5167c3a3-ec3c-4f30-9410-ebc60d61f515-kube-api-access-bddzx\") pod \"community-operators-8ttkg\" (UID: \"5167c3a3-ec3c-4f30-9410-ebc60d61f515\") " pod="openshift-marketplace/community-operators-8ttkg" Jan 26 17:24:28 crc kubenswrapper[4856]: I0126 17:24:28.236845 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5167c3a3-ec3c-4f30-9410-ebc60d61f515-catalog-content\") pod \"community-operators-8ttkg\" (UID: \"5167c3a3-ec3c-4f30-9410-ebc60d61f515\") " pod="openshift-marketplace/community-operators-8ttkg" Jan 26 17:24:28 crc kubenswrapper[4856]: I0126 17:24:28.236893 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5167c3a3-ec3c-4f30-9410-ebc60d61f515-utilities\") pod \"community-operators-8ttkg\" (UID: \"5167c3a3-ec3c-4f30-9410-ebc60d61f515\") " pod="openshift-marketplace/community-operators-8ttkg" Jan 26 17:24:28 crc kubenswrapper[4856]: I0126 17:24:28.338037 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bddzx\" (UniqueName: \"kubernetes.io/projected/5167c3a3-ec3c-4f30-9410-ebc60d61f515-kube-api-access-bddzx\") pod \"community-operators-8ttkg\" (UID: \"5167c3a3-ec3c-4f30-9410-ebc60d61f515\") " pod="openshift-marketplace/community-operators-8ttkg" Jan 26 17:24:28 crc kubenswrapper[4856]: I0126 17:24:28.338112 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5167c3a3-ec3c-4f30-9410-ebc60d61f515-catalog-content\") pod \"community-operators-8ttkg\" (UID: \"5167c3a3-ec3c-4f30-9410-ebc60d61f515\") " pod="openshift-marketplace/community-operators-8ttkg" Jan 26 17:24:28 crc kubenswrapper[4856]: I0126 17:24:28.338171 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5167c3a3-ec3c-4f30-9410-ebc60d61f515-utilities\") pod \"community-operators-8ttkg\" (UID: \"5167c3a3-ec3c-4f30-9410-ebc60d61f515\") " pod="openshift-marketplace/community-operators-8ttkg" Jan 26 17:24:28 crc kubenswrapper[4856]: I0126 17:24:28.338746 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5167c3a3-ec3c-4f30-9410-ebc60d61f515-catalog-content\") pod \"community-operators-8ttkg\" (UID: \"5167c3a3-ec3c-4f30-9410-ebc60d61f515\") " pod="openshift-marketplace/community-operators-8ttkg" Jan 26 17:24:28 crc kubenswrapper[4856]: I0126 17:24:28.338903 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5167c3a3-ec3c-4f30-9410-ebc60d61f515-utilities\") pod \"community-operators-8ttkg\" (UID: \"5167c3a3-ec3c-4f30-9410-ebc60d61f515\") " pod="openshift-marketplace/community-operators-8ttkg" Jan 26 17:24:28 crc kubenswrapper[4856]: I0126 17:24:28.367981 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bddzx\" (UniqueName: \"kubernetes.io/projected/5167c3a3-ec3c-4f30-9410-ebc60d61f515-kube-api-access-bddzx\") pod \"community-operators-8ttkg\" (UID: \"5167c3a3-ec3c-4f30-9410-ebc60d61f515\") " pod="openshift-marketplace/community-operators-8ttkg" Jan 26 17:24:28 crc kubenswrapper[4856]: I0126 17:24:28.465327 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8ttkg" Jan 26 17:24:28 crc kubenswrapper[4856]: I0126 17:24:28.742087 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8ttkg"] Jan 26 17:24:28 crc kubenswrapper[4856]: I0126 17:24:28.855422 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/prometheus-webhook-snmp-4-build"] Jan 26 17:24:28 crc kubenswrapper[4856]: I0126 17:24:28.872893 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-4-build" Jan 26 17:24:28 crc kubenswrapper[4856]: I0126 17:24:28.878480 4856 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-8h4xs" Jan 26 17:24:28 crc kubenswrapper[4856]: I0126 17:24:28.878733 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-4-sys-config" Jan 26 17:24:28 crc kubenswrapper[4856]: I0126 17:24:28.878913 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-4-ca" Jan 26 17:24:28 crc kubenswrapper[4856]: I0126 17:24:28.878969 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-4-global-ca" Jan 26 17:24:28 crc kubenswrapper[4856]: I0126 17:24:28.889266 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-4-build"] Jan 26 17:24:28 crc kubenswrapper[4856]: I0126 17:24:28.913355 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8ttkg" event={"ID":"5167c3a3-ec3c-4f30-9410-ebc60d61f515","Type":"ContainerStarted","Data":"d32dd0fa9bc14484fc546366cd6dea290074e7531292759b6e03e195de0cc5ef"} Jan 26 17:24:28 crc kubenswrapper[4856]: I0126 17:24:28.945361 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/9fd7ae61-20c0-41bb-93e6-f209748133ef-container-storage-run\") pod \"prometheus-webhook-snmp-4-build\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " pod="service-telemetry/prometheus-webhook-snmp-4-build" Jan 26 17:24:28 crc kubenswrapper[4856]: I0126 17:24:28.945432 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cq27\" (UniqueName: \"kubernetes.io/projected/9fd7ae61-20c0-41bb-93e6-f209748133ef-kube-api-access-8cq27\") pod \"prometheus-webhook-snmp-4-build\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " pod="service-telemetry/prometheus-webhook-snmp-4-build" Jan 26 17:24:28 crc kubenswrapper[4856]: I0126 17:24:28.945457 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9fd7ae61-20c0-41bb-93e6-f209748133ef-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-4-build\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " pod="service-telemetry/prometheus-webhook-snmp-4-build" Jan 26 17:24:28 crc kubenswrapper[4856]: I0126 17:24:28.945474 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/9fd7ae61-20c0-41bb-93e6-f209748133ef-container-storage-root\") pod \"prometheus-webhook-snmp-4-build\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " pod="service-telemetry/prometheus-webhook-snmp-4-build" Jan 26 17:24:28 crc kubenswrapper[4856]: I0126 17:24:28.945492 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/9fd7ae61-20c0-41bb-93e6-f209748133ef-buildworkdir\") pod \"prometheus-webhook-snmp-4-build\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " pod="service-telemetry/prometheus-webhook-snmp-4-build" Jan 26 17:24:28 crc kubenswrapper[4856]: I0126 17:24:28.945510 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/9fd7ae61-20c0-41bb-93e6-f209748133ef-build-blob-cache\") pod \"prometheus-webhook-snmp-4-build\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " pod="service-telemetry/prometheus-webhook-snmp-4-build" Jan 26 17:24:28 crc kubenswrapper[4856]: I0126 17:24:28.945542 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/9fd7ae61-20c0-41bb-93e6-f209748133ef-build-system-configs\") pod \"prometheus-webhook-snmp-4-build\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " pod="service-telemetry/prometheus-webhook-snmp-4-build" Jan 26 17:24:28 crc kubenswrapper[4856]: I0126 17:24:28.945569 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/9fd7ae61-20c0-41bb-93e6-f209748133ef-builder-dockercfg-8h4xs-pull\") pod \"prometheus-webhook-snmp-4-build\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " pod="service-telemetry/prometheus-webhook-snmp-4-build" Jan 26 17:24:28 crc kubenswrapper[4856]: I0126 17:24:28.945588 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9fd7ae61-20c0-41bb-93e6-f209748133ef-build-ca-bundles\") pod \"prometheus-webhook-snmp-4-build\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " pod="service-telemetry/prometheus-webhook-snmp-4-build" Jan 26 17:24:28 crc kubenswrapper[4856]: I0126 17:24:28.945625 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9fd7ae61-20c0-41bb-93e6-f209748133ef-node-pullsecrets\") pod \"prometheus-webhook-snmp-4-build\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " pod="service-telemetry/prometheus-webhook-snmp-4-build" Jan 26 17:24:28 crc kubenswrapper[4856]: I0126 17:24:28.945645 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/9fd7ae61-20c0-41bb-93e6-f209748133ef-builder-dockercfg-8h4xs-push\") pod \"prometheus-webhook-snmp-4-build\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " pod="service-telemetry/prometheus-webhook-snmp-4-build" Jan 26 17:24:28 crc kubenswrapper[4856]: I0126 17:24:28.945660 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/9fd7ae61-20c0-41bb-93e6-f209748133ef-buildcachedir\") pod \"prometheus-webhook-snmp-4-build\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " pod="service-telemetry/prometheus-webhook-snmp-4-build" Jan 26 17:24:29 crc kubenswrapper[4856]: I0126 17:24:29.046348 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/9fd7ae61-20c0-41bb-93e6-f209748133ef-buildcachedir\") pod \"prometheus-webhook-snmp-4-build\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " pod="service-telemetry/prometheus-webhook-snmp-4-build" Jan 26 17:24:29 crc kubenswrapper[4856]: I0126 17:24:29.046722 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/9fd7ae61-20c0-41bb-93e6-f209748133ef-builder-dockercfg-8h4xs-push\") pod \"prometheus-webhook-snmp-4-build\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " pod="service-telemetry/prometheus-webhook-snmp-4-build" Jan 26 17:24:29 crc kubenswrapper[4856]: I0126 17:24:29.046501 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/9fd7ae61-20c0-41bb-93e6-f209748133ef-buildcachedir\") pod \"prometheus-webhook-snmp-4-build\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " pod="service-telemetry/prometheus-webhook-snmp-4-build" Jan 26 17:24:29 crc kubenswrapper[4856]: I0126 17:24:29.046755 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/9fd7ae61-20c0-41bb-93e6-f209748133ef-container-storage-run\") pod \"prometheus-webhook-snmp-4-build\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " pod="service-telemetry/prometheus-webhook-snmp-4-build" Jan 26 17:24:29 crc kubenswrapper[4856]: I0126 17:24:29.046901 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cq27\" (UniqueName: \"kubernetes.io/projected/9fd7ae61-20c0-41bb-93e6-f209748133ef-kube-api-access-8cq27\") pod \"prometheus-webhook-snmp-4-build\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " pod="service-telemetry/prometheus-webhook-snmp-4-build" Jan 26 17:24:29 crc kubenswrapper[4856]: I0126 17:24:29.046950 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9fd7ae61-20c0-41bb-93e6-f209748133ef-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-4-build\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " pod="service-telemetry/prometheus-webhook-snmp-4-build" Jan 26 17:24:29 crc kubenswrapper[4856]: I0126 17:24:29.046972 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/9fd7ae61-20c0-41bb-93e6-f209748133ef-container-storage-root\") pod \"prometheus-webhook-snmp-4-build\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " pod="service-telemetry/prometheus-webhook-snmp-4-build" Jan 26 17:24:29 crc kubenswrapper[4856]: I0126 17:24:29.046997 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/9fd7ae61-20c0-41bb-93e6-f209748133ef-buildworkdir\") pod \"prometheus-webhook-snmp-4-build\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " pod="service-telemetry/prometheus-webhook-snmp-4-build" Jan 26 17:24:29 crc kubenswrapper[4856]: I0126 17:24:29.047030 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/9fd7ae61-20c0-41bb-93e6-f209748133ef-build-blob-cache\") pod \"prometheus-webhook-snmp-4-build\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " pod="service-telemetry/prometheus-webhook-snmp-4-build" Jan 26 17:24:29 crc kubenswrapper[4856]: I0126 17:24:29.047058 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/9fd7ae61-20c0-41bb-93e6-f209748133ef-build-system-configs\") pod \"prometheus-webhook-snmp-4-build\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " pod="service-telemetry/prometheus-webhook-snmp-4-build" Jan 26 17:24:29 crc kubenswrapper[4856]: I0126 17:24:29.047101 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/9fd7ae61-20c0-41bb-93e6-f209748133ef-builder-dockercfg-8h4xs-pull\") pod \"prometheus-webhook-snmp-4-build\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " pod="service-telemetry/prometheus-webhook-snmp-4-build" Jan 26 17:24:29 crc kubenswrapper[4856]: I0126 17:24:29.047120 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/9fd7ae61-20c0-41bb-93e6-f209748133ef-container-storage-run\") pod \"prometheus-webhook-snmp-4-build\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " pod="service-telemetry/prometheus-webhook-snmp-4-build" Jan 26 17:24:29 crc kubenswrapper[4856]: I0126 17:24:29.047124 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9fd7ae61-20c0-41bb-93e6-f209748133ef-build-ca-bundles\") pod \"prometheus-webhook-snmp-4-build\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " pod="service-telemetry/prometheus-webhook-snmp-4-build" Jan 26 17:24:29 crc kubenswrapper[4856]: I0126 17:24:29.047188 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9fd7ae61-20c0-41bb-93e6-f209748133ef-node-pullsecrets\") pod \"prometheus-webhook-snmp-4-build\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " pod="service-telemetry/prometheus-webhook-snmp-4-build" Jan 26 17:24:29 crc kubenswrapper[4856]: I0126 17:24:29.047249 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9fd7ae61-20c0-41bb-93e6-f209748133ef-node-pullsecrets\") pod \"prometheus-webhook-snmp-4-build\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " pod="service-telemetry/prometheus-webhook-snmp-4-build" Jan 26 17:24:29 crc kubenswrapper[4856]: I0126 17:24:29.048000 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/9fd7ae61-20c0-41bb-93e6-f209748133ef-buildworkdir\") pod \"prometheus-webhook-snmp-4-build\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " pod="service-telemetry/prometheus-webhook-snmp-4-build" Jan 26 17:24:29 crc kubenswrapper[4856]: I0126 17:24:29.048244 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/9fd7ae61-20c0-41bb-93e6-f209748133ef-build-system-configs\") pod \"prometheus-webhook-snmp-4-build\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " pod="service-telemetry/prometheus-webhook-snmp-4-build" Jan 26 17:24:29 crc kubenswrapper[4856]: I0126 17:24:29.048281 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9fd7ae61-20c0-41bb-93e6-f209748133ef-build-ca-bundles\") pod \"prometheus-webhook-snmp-4-build\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " pod="service-telemetry/prometheus-webhook-snmp-4-build" Jan 26 17:24:29 crc kubenswrapper[4856]: I0126 17:24:29.048383 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9fd7ae61-20c0-41bb-93e6-f209748133ef-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-4-build\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " pod="service-telemetry/prometheus-webhook-snmp-4-build" Jan 26 17:24:29 crc kubenswrapper[4856]: I0126 17:24:29.048453 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/9fd7ae61-20c0-41bb-93e6-f209748133ef-build-blob-cache\") pod \"prometheus-webhook-snmp-4-build\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " pod="service-telemetry/prometheus-webhook-snmp-4-build" Jan 26 17:24:29 crc kubenswrapper[4856]: I0126 17:24:29.048555 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/9fd7ae61-20c0-41bb-93e6-f209748133ef-container-storage-root\") pod \"prometheus-webhook-snmp-4-build\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " pod="service-telemetry/prometheus-webhook-snmp-4-build" Jan 26 17:24:29 crc kubenswrapper[4856]: I0126 17:24:29.061293 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/9fd7ae61-20c0-41bb-93e6-f209748133ef-builder-dockercfg-8h4xs-push\") pod \"prometheus-webhook-snmp-4-build\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " pod="service-telemetry/prometheus-webhook-snmp-4-build" Jan 26 17:24:29 crc kubenswrapper[4856]: I0126 17:24:29.062620 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/9fd7ae61-20c0-41bb-93e6-f209748133ef-builder-dockercfg-8h4xs-pull\") pod \"prometheus-webhook-snmp-4-build\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " pod="service-telemetry/prometheus-webhook-snmp-4-build" Jan 26 17:24:29 crc kubenswrapper[4856]: I0126 17:24:29.066770 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cq27\" (UniqueName: \"kubernetes.io/projected/9fd7ae61-20c0-41bb-93e6-f209748133ef-kube-api-access-8cq27\") pod \"prometheus-webhook-snmp-4-build\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " pod="service-telemetry/prometheus-webhook-snmp-4-build" Jan 26 17:24:29 crc kubenswrapper[4856]: I0126 17:24:29.191650 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-4-build" Jan 26 17:24:29 crc kubenswrapper[4856]: I0126 17:24:29.387757 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-4-build"] Jan 26 17:24:29 crc kubenswrapper[4856]: I0126 17:24:29.923413 4856 generic.go:334] "Generic (PLEG): container finished" podID="5167c3a3-ec3c-4f30-9410-ebc60d61f515" containerID="ea789d382e8ee7680a387d4c0ac031fe25c48fa1835d8928284e77f51f936f6a" exitCode=0 Jan 26 17:24:29 crc kubenswrapper[4856]: I0126 17:24:29.923490 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8ttkg" event={"ID":"5167c3a3-ec3c-4f30-9410-ebc60d61f515","Type":"ContainerDied","Data":"ea789d382e8ee7680a387d4c0ac031fe25c48fa1835d8928284e77f51f936f6a"} Jan 26 17:24:29 crc kubenswrapper[4856]: I0126 17:24:29.926405 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-4-build" event={"ID":"9fd7ae61-20c0-41bb-93e6-f209748133ef","Type":"ContainerStarted","Data":"63ec08bc14f1b2ac4e6ce7fff75e97dfd901f66d0eb16464110e649714bdfefb"} Jan 26 17:24:29 crc kubenswrapper[4856]: I0126 17:24:29.926494 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-4-build" event={"ID":"9fd7ae61-20c0-41bb-93e6-f209748133ef","Type":"ContainerStarted","Data":"1114ab5ebbed8464e1df1efeb4253f6b4ab7782e2e7ad6231ca36374142b6d8b"} Jan 26 17:24:30 crc kubenswrapper[4856]: E0126 17:24:30.051441 4856 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.241:46256->38.102.83.241:41827: write tcp 38.102.83.241:46256->38.102.83.241:41827: write: broken pipe Jan 26 17:24:30 crc kubenswrapper[4856]: I0126 17:24:30.935370 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8ttkg" event={"ID":"5167c3a3-ec3c-4f30-9410-ebc60d61f515","Type":"ContainerStarted","Data":"26c1821c4f23578ac6a0c1415d6730caff4fddaf6502ee746d86ad133657cf42"} Jan 26 17:24:30 crc kubenswrapper[4856]: I0126 17:24:30.938640 4856 generic.go:334] "Generic (PLEG): container finished" podID="9fd7ae61-20c0-41bb-93e6-f209748133ef" containerID="63ec08bc14f1b2ac4e6ce7fff75e97dfd901f66d0eb16464110e649714bdfefb" exitCode=0 Jan 26 17:24:30 crc kubenswrapper[4856]: I0126 17:24:30.938703 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-4-build" event={"ID":"9fd7ae61-20c0-41bb-93e6-f209748133ef","Type":"ContainerDied","Data":"63ec08bc14f1b2ac4e6ce7fff75e97dfd901f66d0eb16464110e649714bdfefb"} Jan 26 17:24:31 crc kubenswrapper[4856]: I0126 17:24:31.946373 4856 generic.go:334] "Generic (PLEG): container finished" podID="5167c3a3-ec3c-4f30-9410-ebc60d61f515" containerID="26c1821c4f23578ac6a0c1415d6730caff4fddaf6502ee746d86ad133657cf42" exitCode=0 Jan 26 17:24:31 crc kubenswrapper[4856]: I0126 17:24:31.946436 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8ttkg" event={"ID":"5167c3a3-ec3c-4f30-9410-ebc60d61f515","Type":"ContainerDied","Data":"26c1821c4f23578ac6a0c1415d6730caff4fddaf6502ee746d86ad133657cf42"} Jan 26 17:24:31 crc kubenswrapper[4856]: I0126 17:24:31.948629 4856 generic.go:334] "Generic (PLEG): container finished" podID="9fd7ae61-20c0-41bb-93e6-f209748133ef" containerID="0af3b5710b654d0464312b2ab2879d5bcc53b56905655509964916d5b57c7f23" exitCode=0 Jan 26 17:24:31 crc kubenswrapper[4856]: I0126 17:24:31.948670 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-4-build" event={"ID":"9fd7ae61-20c0-41bb-93e6-f209748133ef","Type":"ContainerDied","Data":"0af3b5710b654d0464312b2ab2879d5bcc53b56905655509964916d5b57c7f23"} Jan 26 17:24:32 crc kubenswrapper[4856]: I0126 17:24:32.005386 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-webhook-snmp-4-build_9fd7ae61-20c0-41bb-93e6-f209748133ef/manage-dockerfile/0.log" Jan 26 17:24:32 crc kubenswrapper[4856]: I0126 17:24:32.956840 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-4-build" event={"ID":"9fd7ae61-20c0-41bb-93e6-f209748133ef","Type":"ContainerStarted","Data":"02dea6bff540ea6fdb0a80d31de9614f99f56cd175ee20e708f9fb1b894208a9"} Jan 26 17:24:32 crc kubenswrapper[4856]: I0126 17:24:32.988208 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/prometheus-webhook-snmp-4-build" podStartSLOduration=4.988183756 podStartE2EDuration="4.988183756s" podCreationTimestamp="2026-01-26 17:24:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:24:32.982620857 +0000 UTC m=+1568.935874848" watchObservedRunningTime="2026-01-26 17:24:32.988183756 +0000 UTC m=+1568.941437737" Jan 26 17:24:33 crc kubenswrapper[4856]: I0126 17:24:33.966046 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8ttkg" event={"ID":"5167c3a3-ec3c-4f30-9410-ebc60d61f515","Type":"ContainerStarted","Data":"bc8203a7d47ee43722cf1047b5eda9c60a27c27e6ee323351ad4ef557fd2f359"} Jan 26 17:24:33 crc kubenswrapper[4856]: I0126 17:24:33.993128 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8ttkg" podStartSLOduration=2.202157824 podStartE2EDuration="5.993092323s" podCreationTimestamp="2026-01-26 17:24:28 +0000 UTC" firstStartedPulling="2026-01-26 17:24:29.925724982 +0000 UTC m=+1565.878978983" lastFinishedPulling="2026-01-26 17:24:33.716659501 +0000 UTC m=+1569.669913482" observedRunningTime="2026-01-26 17:24:33.986826944 +0000 UTC m=+1569.940080945" watchObservedRunningTime="2026-01-26 17:24:33.993092323 +0000 UTC m=+1569.946346354" Jan 26 17:24:38 crc kubenswrapper[4856]: I0126 17:24:38.466569 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8ttkg" Jan 26 17:24:38 crc kubenswrapper[4856]: I0126 17:24:38.467073 4856 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8ttkg" Jan 26 17:24:38 crc kubenswrapper[4856]: I0126 17:24:38.524489 4856 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8ttkg" Jan 26 17:24:39 crc kubenswrapper[4856]: I0126 17:24:39.038001 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8ttkg" Jan 26 17:24:39 crc kubenswrapper[4856]: I0126 17:24:39.096842 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8ttkg"] Jan 26 17:24:41 crc kubenswrapper[4856]: I0126 17:24:41.057984 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8ttkg" podUID="5167c3a3-ec3c-4f30-9410-ebc60d61f515" containerName="registry-server" containerID="cri-o://bc8203a7d47ee43722cf1047b5eda9c60a27c27e6ee323351ad4ef557fd2f359" gracePeriod=2 Jan 26 17:24:44 crc kubenswrapper[4856]: I0126 17:24:44.086723 4856 generic.go:334] "Generic (PLEG): container finished" podID="5167c3a3-ec3c-4f30-9410-ebc60d61f515" containerID="bc8203a7d47ee43722cf1047b5eda9c60a27c27e6ee323351ad4ef557fd2f359" exitCode=0 Jan 26 17:24:44 crc kubenswrapper[4856]: I0126 17:24:44.086784 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8ttkg" event={"ID":"5167c3a3-ec3c-4f30-9410-ebc60d61f515","Type":"ContainerDied","Data":"bc8203a7d47ee43722cf1047b5eda9c60a27c27e6ee323351ad4ef557fd2f359"} Jan 26 17:24:44 crc kubenswrapper[4856]: I0126 17:24:44.087207 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8ttkg" event={"ID":"5167c3a3-ec3c-4f30-9410-ebc60d61f515","Type":"ContainerDied","Data":"d32dd0fa9bc14484fc546366cd6dea290074e7531292759b6e03e195de0cc5ef"} Jan 26 17:24:44 crc kubenswrapper[4856]: I0126 17:24:44.087229 4856 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d32dd0fa9bc14484fc546366cd6dea290074e7531292759b6e03e195de0cc5ef" Jan 26 17:24:44 crc kubenswrapper[4856]: I0126 17:24:44.098656 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8ttkg" Jan 26 17:24:44 crc kubenswrapper[4856]: I0126 17:24:44.208873 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5167c3a3-ec3c-4f30-9410-ebc60d61f515-catalog-content\") pod \"5167c3a3-ec3c-4f30-9410-ebc60d61f515\" (UID: \"5167c3a3-ec3c-4f30-9410-ebc60d61f515\") " Jan 26 17:24:44 crc kubenswrapper[4856]: I0126 17:24:44.209107 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5167c3a3-ec3c-4f30-9410-ebc60d61f515-utilities\") pod \"5167c3a3-ec3c-4f30-9410-ebc60d61f515\" (UID: \"5167c3a3-ec3c-4f30-9410-ebc60d61f515\") " Jan 26 17:24:44 crc kubenswrapper[4856]: I0126 17:24:44.209197 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bddzx\" (UniqueName: \"kubernetes.io/projected/5167c3a3-ec3c-4f30-9410-ebc60d61f515-kube-api-access-bddzx\") pod \"5167c3a3-ec3c-4f30-9410-ebc60d61f515\" (UID: \"5167c3a3-ec3c-4f30-9410-ebc60d61f515\") " Jan 26 17:24:44 crc kubenswrapper[4856]: I0126 17:24:44.209928 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5167c3a3-ec3c-4f30-9410-ebc60d61f515-utilities" (OuterVolumeSpecName: "utilities") pod "5167c3a3-ec3c-4f30-9410-ebc60d61f515" (UID: "5167c3a3-ec3c-4f30-9410-ebc60d61f515"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:24:44 crc kubenswrapper[4856]: I0126 17:24:44.214998 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5167c3a3-ec3c-4f30-9410-ebc60d61f515-kube-api-access-bddzx" (OuterVolumeSpecName: "kube-api-access-bddzx") pod "5167c3a3-ec3c-4f30-9410-ebc60d61f515" (UID: "5167c3a3-ec3c-4f30-9410-ebc60d61f515"). InnerVolumeSpecName "kube-api-access-bddzx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:24:44 crc kubenswrapper[4856]: I0126 17:24:44.282887 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5167c3a3-ec3c-4f30-9410-ebc60d61f515-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5167c3a3-ec3c-4f30-9410-ebc60d61f515" (UID: "5167c3a3-ec3c-4f30-9410-ebc60d61f515"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:24:44 crc kubenswrapper[4856]: I0126 17:24:44.311144 4856 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5167c3a3-ec3c-4f30-9410-ebc60d61f515-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 17:24:44 crc kubenswrapper[4856]: I0126 17:24:44.311190 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bddzx\" (UniqueName: \"kubernetes.io/projected/5167c3a3-ec3c-4f30-9410-ebc60d61f515-kube-api-access-bddzx\") on node \"crc\" DevicePath \"\"" Jan 26 17:24:44 crc kubenswrapper[4856]: I0126 17:24:44.311205 4856 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5167c3a3-ec3c-4f30-9410-ebc60d61f515-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 17:24:45 crc kubenswrapper[4856]: I0126 17:24:45.094298 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8ttkg" Jan 26 17:24:45 crc kubenswrapper[4856]: I0126 17:24:45.137909 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8ttkg"] Jan 26 17:24:45 crc kubenswrapper[4856]: I0126 17:24:45.142976 4856 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8ttkg"] Jan 26 17:24:45 crc kubenswrapper[4856]: I0126 17:24:45.402902 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5167c3a3-ec3c-4f30-9410-ebc60d61f515" path="/var/lib/kubelet/pods/5167c3a3-ec3c-4f30-9410-ebc60d61f515/volumes" Jan 26 17:25:28 crc kubenswrapper[4856]: I0126 17:25:28.510993 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-webhook-snmp-4-build_9fd7ae61-20c0-41bb-93e6-f209748133ef/docker-build/0.log" Jan 26 17:25:28 crc kubenswrapper[4856]: I0126 17:25:28.512334 4856 generic.go:334] "Generic (PLEG): container finished" podID="9fd7ae61-20c0-41bb-93e6-f209748133ef" containerID="02dea6bff540ea6fdb0a80d31de9614f99f56cd175ee20e708f9fb1b894208a9" exitCode=1 Jan 26 17:25:28 crc kubenswrapper[4856]: I0126 17:25:28.512369 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-4-build" event={"ID":"9fd7ae61-20c0-41bb-93e6-f209748133ef","Type":"ContainerDied","Data":"02dea6bff540ea6fdb0a80d31de9614f99f56cd175ee20e708f9fb1b894208a9"} Jan 26 17:25:29 crc kubenswrapper[4856]: I0126 17:25:29.782450 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-webhook-snmp-4-build_9fd7ae61-20c0-41bb-93e6-f209748133ef/docker-build/0.log" Jan 26 17:25:29 crc kubenswrapper[4856]: I0126 17:25:29.785541 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-4-build" Jan 26 17:25:29 crc kubenswrapper[4856]: I0126 17:25:29.973339 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/9fd7ae61-20c0-41bb-93e6-f209748133ef-builder-dockercfg-8h4xs-push\") pod \"9fd7ae61-20c0-41bb-93e6-f209748133ef\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " Jan 26 17:25:29 crc kubenswrapper[4856]: I0126 17:25:29.973424 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/9fd7ae61-20c0-41bb-93e6-f209748133ef-buildcachedir\") pod \"9fd7ae61-20c0-41bb-93e6-f209748133ef\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " Jan 26 17:25:29 crc kubenswrapper[4856]: I0126 17:25:29.973472 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9fd7ae61-20c0-41bb-93e6-f209748133ef-build-proxy-ca-bundles\") pod \"9fd7ae61-20c0-41bb-93e6-f209748133ef\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " Jan 26 17:25:29 crc kubenswrapper[4856]: I0126 17:25:29.973503 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/9fd7ae61-20c0-41bb-93e6-f209748133ef-builder-dockercfg-8h4xs-pull\") pod \"9fd7ae61-20c0-41bb-93e6-f209748133ef\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " Jan 26 17:25:29 crc kubenswrapper[4856]: I0126 17:25:29.973556 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9fd7ae61-20c0-41bb-93e6-f209748133ef-node-pullsecrets\") pod \"9fd7ae61-20c0-41bb-93e6-f209748133ef\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " Jan 26 17:25:29 crc kubenswrapper[4856]: I0126 17:25:29.973621 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cq27\" (UniqueName: \"kubernetes.io/projected/9fd7ae61-20c0-41bb-93e6-f209748133ef-kube-api-access-8cq27\") pod \"9fd7ae61-20c0-41bb-93e6-f209748133ef\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " Jan 26 17:25:29 crc kubenswrapper[4856]: I0126 17:25:29.973675 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/9fd7ae61-20c0-41bb-93e6-f209748133ef-build-blob-cache\") pod \"9fd7ae61-20c0-41bb-93e6-f209748133ef\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " Jan 26 17:25:29 crc kubenswrapper[4856]: I0126 17:25:29.973703 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9fd7ae61-20c0-41bb-93e6-f209748133ef-build-ca-bundles\") pod \"9fd7ae61-20c0-41bb-93e6-f209748133ef\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " Jan 26 17:25:29 crc kubenswrapper[4856]: I0126 17:25:29.973785 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/9fd7ae61-20c0-41bb-93e6-f209748133ef-container-storage-root\") pod \"9fd7ae61-20c0-41bb-93e6-f209748133ef\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " Jan 26 17:25:29 crc kubenswrapper[4856]: I0126 17:25:29.973821 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/9fd7ae61-20c0-41bb-93e6-f209748133ef-build-system-configs\") pod \"9fd7ae61-20c0-41bb-93e6-f209748133ef\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " Jan 26 17:25:29 crc kubenswrapper[4856]: I0126 17:25:29.974066 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9fd7ae61-20c0-41bb-93e6-f209748133ef-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "9fd7ae61-20c0-41bb-93e6-f209748133ef" (UID: "9fd7ae61-20c0-41bb-93e6-f209748133ef"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:25:29 crc kubenswrapper[4856]: I0126 17:25:29.974614 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9fd7ae61-20c0-41bb-93e6-f209748133ef-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "9fd7ae61-20c0-41bb-93e6-f209748133ef" (UID: "9fd7ae61-20c0-41bb-93e6-f209748133ef"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:25:29 crc kubenswrapper[4856]: I0126 17:25:29.974953 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9fd7ae61-20c0-41bb-93e6-f209748133ef-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "9fd7ae61-20c0-41bb-93e6-f209748133ef" (UID: "9fd7ae61-20c0-41bb-93e6-f209748133ef"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:25:29 crc kubenswrapper[4856]: I0126 17:25:29.975029 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/9fd7ae61-20c0-41bb-93e6-f209748133ef-container-storage-run\") pod \"9fd7ae61-20c0-41bb-93e6-f209748133ef\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " Jan 26 17:25:29 crc kubenswrapper[4856]: I0126 17:25:29.975459 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/9fd7ae61-20c0-41bb-93e6-f209748133ef-buildworkdir\") pod \"9fd7ae61-20c0-41bb-93e6-f209748133ef\" (UID: \"9fd7ae61-20c0-41bb-93e6-f209748133ef\") " Jan 26 17:25:29 crc kubenswrapper[4856]: I0126 17:25:29.975033 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9fd7ae61-20c0-41bb-93e6-f209748133ef-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "9fd7ae61-20c0-41bb-93e6-f209748133ef" (UID: "9fd7ae61-20c0-41bb-93e6-f209748133ef"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:25:29 crc kubenswrapper[4856]: I0126 17:25:29.975361 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9fd7ae61-20c0-41bb-93e6-f209748133ef-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "9fd7ae61-20c0-41bb-93e6-f209748133ef" (UID: "9fd7ae61-20c0-41bb-93e6-f209748133ef"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:25:29 crc kubenswrapper[4856]: I0126 17:25:29.975501 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9fd7ae61-20c0-41bb-93e6-f209748133ef-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "9fd7ae61-20c0-41bb-93e6-f209748133ef" (UID: "9fd7ae61-20c0-41bb-93e6-f209748133ef"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:25:29 crc kubenswrapper[4856]: I0126 17:25:29.975990 4856 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9fd7ae61-20c0-41bb-93e6-f209748133ef-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 26 17:25:29 crc kubenswrapper[4856]: I0126 17:25:29.976033 4856 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/9fd7ae61-20c0-41bb-93e6-f209748133ef-build-system-configs\") on node \"crc\" DevicePath \"\"" Jan 26 17:25:29 crc kubenswrapper[4856]: I0126 17:25:29.976054 4856 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/9fd7ae61-20c0-41bb-93e6-f209748133ef-container-storage-run\") on node \"crc\" DevicePath \"\"" Jan 26 17:25:29 crc kubenswrapper[4856]: I0126 17:25:29.976072 4856 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/9fd7ae61-20c0-41bb-93e6-f209748133ef-buildcachedir\") on node \"crc\" DevicePath \"\"" Jan 26 17:25:29 crc kubenswrapper[4856]: I0126 17:25:29.976088 4856 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9fd7ae61-20c0-41bb-93e6-f209748133ef-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 26 17:25:29 crc kubenswrapper[4856]: I0126 17:25:29.976105 4856 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9fd7ae61-20c0-41bb-93e6-f209748133ef-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Jan 26 17:25:29 crc kubenswrapper[4856]: I0126 17:25:29.977926 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9fd7ae61-20c0-41bb-93e6-f209748133ef-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "9fd7ae61-20c0-41bb-93e6-f209748133ef" (UID: "9fd7ae61-20c0-41bb-93e6-f209748133ef"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:25:29 crc kubenswrapper[4856]: I0126 17:25:29.980418 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fd7ae61-20c0-41bb-93e6-f209748133ef-builder-dockercfg-8h4xs-push" (OuterVolumeSpecName: "builder-dockercfg-8h4xs-push") pod "9fd7ae61-20c0-41bb-93e6-f209748133ef" (UID: "9fd7ae61-20c0-41bb-93e6-f209748133ef"). InnerVolumeSpecName "builder-dockercfg-8h4xs-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:25:29 crc kubenswrapper[4856]: I0126 17:25:29.980543 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fd7ae61-20c0-41bb-93e6-f209748133ef-kube-api-access-8cq27" (OuterVolumeSpecName: "kube-api-access-8cq27") pod "9fd7ae61-20c0-41bb-93e6-f209748133ef" (UID: "9fd7ae61-20c0-41bb-93e6-f209748133ef"). InnerVolumeSpecName "kube-api-access-8cq27". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:25:29 crc kubenswrapper[4856]: I0126 17:25:29.981039 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fd7ae61-20c0-41bb-93e6-f209748133ef-builder-dockercfg-8h4xs-pull" (OuterVolumeSpecName: "builder-dockercfg-8h4xs-pull") pod "9fd7ae61-20c0-41bb-93e6-f209748133ef" (UID: "9fd7ae61-20c0-41bb-93e6-f209748133ef"). InnerVolumeSpecName "builder-dockercfg-8h4xs-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:25:30 crc kubenswrapper[4856]: I0126 17:25:30.060554 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9fd7ae61-20c0-41bb-93e6-f209748133ef-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "9fd7ae61-20c0-41bb-93e6-f209748133ef" (UID: "9fd7ae61-20c0-41bb-93e6-f209748133ef"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:25:30 crc kubenswrapper[4856]: I0126 17:25:30.076993 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cq27\" (UniqueName: \"kubernetes.io/projected/9fd7ae61-20c0-41bb-93e6-f209748133ef-kube-api-access-8cq27\") on node \"crc\" DevicePath \"\"" Jan 26 17:25:30 crc kubenswrapper[4856]: I0126 17:25:30.077028 4856 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/9fd7ae61-20c0-41bb-93e6-f209748133ef-build-blob-cache\") on node \"crc\" DevicePath \"\"" Jan 26 17:25:30 crc kubenswrapper[4856]: I0126 17:25:30.077041 4856 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/9fd7ae61-20c0-41bb-93e6-f209748133ef-buildworkdir\") on node \"crc\" DevicePath \"\"" Jan 26 17:25:30 crc kubenswrapper[4856]: I0126 17:25:30.077053 4856 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/9fd7ae61-20c0-41bb-93e6-f209748133ef-builder-dockercfg-8h4xs-push\") on node \"crc\" DevicePath \"\"" Jan 26 17:25:30 crc kubenswrapper[4856]: I0126 17:25:30.077066 4856 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/9fd7ae61-20c0-41bb-93e6-f209748133ef-builder-dockercfg-8h4xs-pull\") on node \"crc\" DevicePath \"\"" Jan 26 17:25:30 crc kubenswrapper[4856]: I0126 17:25:30.530510 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-webhook-snmp-4-build_9fd7ae61-20c0-41bb-93e6-f209748133ef/docker-build/0.log" Jan 26 17:25:30 crc kubenswrapper[4856]: I0126 17:25:30.531861 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-4-build" event={"ID":"9fd7ae61-20c0-41bb-93e6-f209748133ef","Type":"ContainerDied","Data":"1114ab5ebbed8464e1df1efeb4253f6b4ab7782e2e7ad6231ca36374142b6d8b"} Jan 26 17:25:30 crc kubenswrapper[4856]: I0126 17:25:30.531908 4856 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1114ab5ebbed8464e1df1efeb4253f6b4ab7782e2e7ad6231ca36374142b6d8b" Jan 26 17:25:30 crc kubenswrapper[4856]: I0126 17:25:30.531948 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-4-build" Jan 26 17:25:30 crc kubenswrapper[4856]: I0126 17:25:30.608249 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9fd7ae61-20c0-41bb-93e6-f209748133ef-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "9fd7ae61-20c0-41bb-93e6-f209748133ef" (UID: "9fd7ae61-20c0-41bb-93e6-f209748133ef"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:25:30 crc kubenswrapper[4856]: I0126 17:25:30.687613 4856 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/9fd7ae61-20c0-41bb-93e6-f209748133ef-container-storage-root\") on node \"crc\" DevicePath \"\"" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.145750 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/prometheus-webhook-snmp-5-build"] Jan 26 17:25:40 crc kubenswrapper[4856]: E0126 17:25:40.146589 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fd7ae61-20c0-41bb-93e6-f209748133ef" containerName="manage-dockerfile" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.146606 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fd7ae61-20c0-41bb-93e6-f209748133ef" containerName="manage-dockerfile" Jan 26 17:25:40 crc kubenswrapper[4856]: E0126 17:25:40.146620 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fd7ae61-20c0-41bb-93e6-f209748133ef" containerName="git-clone" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.146626 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fd7ae61-20c0-41bb-93e6-f209748133ef" containerName="git-clone" Jan 26 17:25:40 crc kubenswrapper[4856]: E0126 17:25:40.146635 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5167c3a3-ec3c-4f30-9410-ebc60d61f515" containerName="extract-content" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.146641 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="5167c3a3-ec3c-4f30-9410-ebc60d61f515" containerName="extract-content" Jan 26 17:25:40 crc kubenswrapper[4856]: E0126 17:25:40.146651 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5167c3a3-ec3c-4f30-9410-ebc60d61f515" containerName="registry-server" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.146658 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="5167c3a3-ec3c-4f30-9410-ebc60d61f515" containerName="registry-server" Jan 26 17:25:40 crc kubenswrapper[4856]: E0126 17:25:40.146668 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5167c3a3-ec3c-4f30-9410-ebc60d61f515" containerName="extract-utilities" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.146673 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="5167c3a3-ec3c-4f30-9410-ebc60d61f515" containerName="extract-utilities" Jan 26 17:25:40 crc kubenswrapper[4856]: E0126 17:25:40.146682 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fd7ae61-20c0-41bb-93e6-f209748133ef" containerName="docker-build" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.146689 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fd7ae61-20c0-41bb-93e6-f209748133ef" containerName="docker-build" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.146877 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="5167c3a3-ec3c-4f30-9410-ebc60d61f515" containerName="registry-server" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.146894 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fd7ae61-20c0-41bb-93e6-f209748133ef" containerName="docker-build" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.148023 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-5-build" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.150129 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-5-global-ca" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.150163 4856 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-8h4xs" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.150129 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-5-ca" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.150467 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-5-sys-config" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.164768 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-5-build"] Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.377321 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/77516bf5-33dc-4722-85d1-8cd196e055b3-buildworkdir\") pod \"prometheus-webhook-snmp-5-build\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " pod="service-telemetry/prometheus-webhook-snmp-5-build" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.377474 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/77516bf5-33dc-4722-85d1-8cd196e055b3-node-pullsecrets\") pod \"prometheus-webhook-snmp-5-build\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " pod="service-telemetry/prometheus-webhook-snmp-5-build" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.377558 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/77516bf5-33dc-4722-85d1-8cd196e055b3-builder-dockercfg-8h4xs-pull\") pod \"prometheus-webhook-snmp-5-build\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " pod="service-telemetry/prometheus-webhook-snmp-5-build" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.377603 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/77516bf5-33dc-4722-85d1-8cd196e055b3-build-system-configs\") pod \"prometheus-webhook-snmp-5-build\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " pod="service-telemetry/prometheus-webhook-snmp-5-build" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.377842 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/77516bf5-33dc-4722-85d1-8cd196e055b3-container-storage-run\") pod \"prometheus-webhook-snmp-5-build\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " pod="service-telemetry/prometheus-webhook-snmp-5-build" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.377897 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/77516bf5-33dc-4722-85d1-8cd196e055b3-build-blob-cache\") pod \"prometheus-webhook-snmp-5-build\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " pod="service-telemetry/prometheus-webhook-snmp-5-build" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.377935 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vb88j\" (UniqueName: \"kubernetes.io/projected/77516bf5-33dc-4722-85d1-8cd196e055b3-kube-api-access-vb88j\") pod \"prometheus-webhook-snmp-5-build\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " pod="service-telemetry/prometheus-webhook-snmp-5-build" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.377983 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/77516bf5-33dc-4722-85d1-8cd196e055b3-build-ca-bundles\") pod \"prometheus-webhook-snmp-5-build\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " pod="service-telemetry/prometheus-webhook-snmp-5-build" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.378068 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/77516bf5-33dc-4722-85d1-8cd196e055b3-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-5-build\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " pod="service-telemetry/prometheus-webhook-snmp-5-build" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.378084 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/77516bf5-33dc-4722-85d1-8cd196e055b3-buildcachedir\") pod \"prometheus-webhook-snmp-5-build\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " pod="service-telemetry/prometheus-webhook-snmp-5-build" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.378103 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/77516bf5-33dc-4722-85d1-8cd196e055b3-builder-dockercfg-8h4xs-push\") pod \"prometheus-webhook-snmp-5-build\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " pod="service-telemetry/prometheus-webhook-snmp-5-build" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.378127 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/77516bf5-33dc-4722-85d1-8cd196e055b3-container-storage-root\") pod \"prometheus-webhook-snmp-5-build\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " pod="service-telemetry/prometheus-webhook-snmp-5-build" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.479723 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/77516bf5-33dc-4722-85d1-8cd196e055b3-build-blob-cache\") pod \"prometheus-webhook-snmp-5-build\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " pod="service-telemetry/prometheus-webhook-snmp-5-build" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.479794 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vb88j\" (UniqueName: \"kubernetes.io/projected/77516bf5-33dc-4722-85d1-8cd196e055b3-kube-api-access-vb88j\") pod \"prometheus-webhook-snmp-5-build\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " pod="service-telemetry/prometheus-webhook-snmp-5-build" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.479850 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/77516bf5-33dc-4722-85d1-8cd196e055b3-build-ca-bundles\") pod \"prometheus-webhook-snmp-5-build\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " pod="service-telemetry/prometheus-webhook-snmp-5-build" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.479983 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/77516bf5-33dc-4722-85d1-8cd196e055b3-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-5-build\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " pod="service-telemetry/prometheus-webhook-snmp-5-build" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.480037 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/77516bf5-33dc-4722-85d1-8cd196e055b3-buildcachedir\") pod \"prometheus-webhook-snmp-5-build\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " pod="service-telemetry/prometheus-webhook-snmp-5-build" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.480074 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/77516bf5-33dc-4722-85d1-8cd196e055b3-builder-dockercfg-8h4xs-push\") pod \"prometheus-webhook-snmp-5-build\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " pod="service-telemetry/prometheus-webhook-snmp-5-build" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.480097 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/77516bf5-33dc-4722-85d1-8cd196e055b3-container-storage-root\") pod \"prometheus-webhook-snmp-5-build\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " pod="service-telemetry/prometheus-webhook-snmp-5-build" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.480124 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/77516bf5-33dc-4722-85d1-8cd196e055b3-buildworkdir\") pod \"prometheus-webhook-snmp-5-build\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " pod="service-telemetry/prometheus-webhook-snmp-5-build" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.480183 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/77516bf5-33dc-4722-85d1-8cd196e055b3-node-pullsecrets\") pod \"prometheus-webhook-snmp-5-build\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " pod="service-telemetry/prometheus-webhook-snmp-5-build" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.480220 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/77516bf5-33dc-4722-85d1-8cd196e055b3-builder-dockercfg-8h4xs-pull\") pod \"prometheus-webhook-snmp-5-build\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " pod="service-telemetry/prometheus-webhook-snmp-5-build" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.480248 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/77516bf5-33dc-4722-85d1-8cd196e055b3-build-system-configs\") pod \"prometheus-webhook-snmp-5-build\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " pod="service-telemetry/prometheus-webhook-snmp-5-build" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.480272 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/77516bf5-33dc-4722-85d1-8cd196e055b3-container-storage-run\") pod \"prometheus-webhook-snmp-5-build\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " pod="service-telemetry/prometheus-webhook-snmp-5-build" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.480582 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/77516bf5-33dc-4722-85d1-8cd196e055b3-node-pullsecrets\") pod \"prometheus-webhook-snmp-5-build\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " pod="service-telemetry/prometheus-webhook-snmp-5-build" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.480612 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/77516bf5-33dc-4722-85d1-8cd196e055b3-build-blob-cache\") pod \"prometheus-webhook-snmp-5-build\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " pod="service-telemetry/prometheus-webhook-snmp-5-build" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.480711 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/77516bf5-33dc-4722-85d1-8cd196e055b3-container-storage-run\") pod \"prometheus-webhook-snmp-5-build\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " pod="service-telemetry/prometheus-webhook-snmp-5-build" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.480864 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/77516bf5-33dc-4722-85d1-8cd196e055b3-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-5-build\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " pod="service-telemetry/prometheus-webhook-snmp-5-build" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.480901 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/77516bf5-33dc-4722-85d1-8cd196e055b3-buildcachedir\") pod \"prometheus-webhook-snmp-5-build\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " pod="service-telemetry/prometheus-webhook-snmp-5-build" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.481003 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/77516bf5-33dc-4722-85d1-8cd196e055b3-buildworkdir\") pod \"prometheus-webhook-snmp-5-build\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " pod="service-telemetry/prometheus-webhook-snmp-5-build" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.481210 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/77516bf5-33dc-4722-85d1-8cd196e055b3-container-storage-root\") pod \"prometheus-webhook-snmp-5-build\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " pod="service-telemetry/prometheus-webhook-snmp-5-build" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.481415 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/77516bf5-33dc-4722-85d1-8cd196e055b3-build-system-configs\") pod \"prometheus-webhook-snmp-5-build\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " pod="service-telemetry/prometheus-webhook-snmp-5-build" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.482384 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/77516bf5-33dc-4722-85d1-8cd196e055b3-build-ca-bundles\") pod \"prometheus-webhook-snmp-5-build\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " pod="service-telemetry/prometheus-webhook-snmp-5-build" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.487068 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/77516bf5-33dc-4722-85d1-8cd196e055b3-builder-dockercfg-8h4xs-pull\") pod \"prometheus-webhook-snmp-5-build\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " pod="service-telemetry/prometheus-webhook-snmp-5-build" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.487482 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/77516bf5-33dc-4722-85d1-8cd196e055b3-builder-dockercfg-8h4xs-push\") pod \"prometheus-webhook-snmp-5-build\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " pod="service-telemetry/prometheus-webhook-snmp-5-build" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.506890 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vb88j\" (UniqueName: \"kubernetes.io/projected/77516bf5-33dc-4722-85d1-8cd196e055b3-kube-api-access-vb88j\") pod \"prometheus-webhook-snmp-5-build\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " pod="service-telemetry/prometheus-webhook-snmp-5-build" Jan 26 17:25:40 crc kubenswrapper[4856]: I0126 17:25:40.769464 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-5-build" Jan 26 17:25:41 crc kubenswrapper[4856]: I0126 17:25:41.248592 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-5-build"] Jan 26 17:25:41 crc kubenswrapper[4856]: I0126 17:25:41.620558 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-5-build" event={"ID":"77516bf5-33dc-4722-85d1-8cd196e055b3","Type":"ContainerStarted","Data":"4bf608fd02a4ba9adf66de6dc6325d52af81af1b294ec61636551feb5cd91850"} Jan 26 17:25:41 crc kubenswrapper[4856]: I0126 17:25:41.620612 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-5-build" event={"ID":"77516bf5-33dc-4722-85d1-8cd196e055b3","Type":"ContainerStarted","Data":"73566cebe45397c2722640fc7ffd2e04eb8a444fc2c63f9f96879b8d7aec440b"} Jan 26 17:25:42 crc kubenswrapper[4856]: I0126 17:25:42.629454 4856 generic.go:334] "Generic (PLEG): container finished" podID="77516bf5-33dc-4722-85d1-8cd196e055b3" containerID="4bf608fd02a4ba9adf66de6dc6325d52af81af1b294ec61636551feb5cd91850" exitCode=0 Jan 26 17:25:42 crc kubenswrapper[4856]: I0126 17:25:42.629577 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-5-build" event={"ID":"77516bf5-33dc-4722-85d1-8cd196e055b3","Type":"ContainerDied","Data":"4bf608fd02a4ba9adf66de6dc6325d52af81af1b294ec61636551feb5cd91850"} Jan 26 17:25:43 crc kubenswrapper[4856]: I0126 17:25:43.637489 4856 generic.go:334] "Generic (PLEG): container finished" podID="77516bf5-33dc-4722-85d1-8cd196e055b3" containerID="77d3e37cbe082f8e8cf8830aed97e12cbb79797a661886cf41da5c29c802032d" exitCode=0 Jan 26 17:25:43 crc kubenswrapper[4856]: I0126 17:25:43.637567 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-5-build" event={"ID":"77516bf5-33dc-4722-85d1-8cd196e055b3","Type":"ContainerDied","Data":"77d3e37cbe082f8e8cf8830aed97e12cbb79797a661886cf41da5c29c802032d"} Jan 26 17:25:43 crc kubenswrapper[4856]: I0126 17:25:43.823453 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-webhook-snmp-5-build_77516bf5-33dc-4722-85d1-8cd196e055b3/manage-dockerfile/0.log" Jan 26 17:25:44 crc kubenswrapper[4856]: I0126 17:25:44.647751 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-5-build" event={"ID":"77516bf5-33dc-4722-85d1-8cd196e055b3","Type":"ContainerStarted","Data":"085cbd8b2c148470c95639832aef6196db9721aff2d0bb845f95a05fa42d58e1"} Jan 26 17:25:44 crc kubenswrapper[4856]: I0126 17:25:44.692550 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/prometheus-webhook-snmp-5-build" podStartSLOduration=4.692501115 podStartE2EDuration="4.692501115s" podCreationTimestamp="2026-01-26 17:25:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 17:25:44.687716118 +0000 UTC m=+1640.640970119" watchObservedRunningTime="2026-01-26 17:25:44.692501115 +0000 UTC m=+1640.645755096" Jan 26 17:25:56 crc kubenswrapper[4856]: I0126 17:25:56.938347 4856 patch_prober.go:28] interesting pod/machine-config-daemon-xm9cq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:25:56 crc kubenswrapper[4856]: I0126 17:25:56.938861 4856 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:26:13 crc kubenswrapper[4856]: I0126 17:26:13.863440 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-webhook-snmp-5-build_77516bf5-33dc-4722-85d1-8cd196e055b3/docker-build/0.log" Jan 26 17:26:13 crc kubenswrapper[4856]: I0126 17:26:13.865870 4856 generic.go:334] "Generic (PLEG): container finished" podID="77516bf5-33dc-4722-85d1-8cd196e055b3" containerID="085cbd8b2c148470c95639832aef6196db9721aff2d0bb845f95a05fa42d58e1" exitCode=1 Jan 26 17:26:13 crc kubenswrapper[4856]: I0126 17:26:13.865919 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-5-build" event={"ID":"77516bf5-33dc-4722-85d1-8cd196e055b3","Type":"ContainerDied","Data":"085cbd8b2c148470c95639832aef6196db9721aff2d0bb845f95a05fa42d58e1"} Jan 26 17:26:15 crc kubenswrapper[4856]: I0126 17:26:15.120368 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-webhook-snmp-5-build_77516bf5-33dc-4722-85d1-8cd196e055b3/docker-build/0.log" Jan 26 17:26:15 crc kubenswrapper[4856]: I0126 17:26:15.121771 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-5-build" Jan 26 17:26:15 crc kubenswrapper[4856]: I0126 17:26:15.248832 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/77516bf5-33dc-4722-85d1-8cd196e055b3-builder-dockercfg-8h4xs-pull\") pod \"77516bf5-33dc-4722-85d1-8cd196e055b3\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " Jan 26 17:26:15 crc kubenswrapper[4856]: I0126 17:26:15.248891 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/77516bf5-33dc-4722-85d1-8cd196e055b3-node-pullsecrets\") pod \"77516bf5-33dc-4722-85d1-8cd196e055b3\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " Jan 26 17:26:15 crc kubenswrapper[4856]: I0126 17:26:15.248935 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/77516bf5-33dc-4722-85d1-8cd196e055b3-builder-dockercfg-8h4xs-push\") pod \"77516bf5-33dc-4722-85d1-8cd196e055b3\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " Jan 26 17:26:15 crc kubenswrapper[4856]: I0126 17:26:15.248965 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/77516bf5-33dc-4722-85d1-8cd196e055b3-build-system-configs\") pod \"77516bf5-33dc-4722-85d1-8cd196e055b3\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " Jan 26 17:26:15 crc kubenswrapper[4856]: I0126 17:26:15.248996 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/77516bf5-33dc-4722-85d1-8cd196e055b3-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "77516bf5-33dc-4722-85d1-8cd196e055b3" (UID: "77516bf5-33dc-4722-85d1-8cd196e055b3"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:26:15 crc kubenswrapper[4856]: I0126 17:26:15.249020 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/77516bf5-33dc-4722-85d1-8cd196e055b3-container-storage-run\") pod \"77516bf5-33dc-4722-85d1-8cd196e055b3\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " Jan 26 17:26:15 crc kubenswrapper[4856]: I0126 17:26:15.249158 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/77516bf5-33dc-4722-85d1-8cd196e055b3-build-blob-cache\") pod \"77516bf5-33dc-4722-85d1-8cd196e055b3\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " Jan 26 17:26:15 crc kubenswrapper[4856]: I0126 17:26:15.249190 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/77516bf5-33dc-4722-85d1-8cd196e055b3-build-ca-bundles\") pod \"77516bf5-33dc-4722-85d1-8cd196e055b3\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " Jan 26 17:26:15 crc kubenswrapper[4856]: I0126 17:26:15.249214 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/77516bf5-33dc-4722-85d1-8cd196e055b3-buildworkdir\") pod \"77516bf5-33dc-4722-85d1-8cd196e055b3\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " Jan 26 17:26:15 crc kubenswrapper[4856]: I0126 17:26:15.249236 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/77516bf5-33dc-4722-85d1-8cd196e055b3-build-proxy-ca-bundles\") pod \"77516bf5-33dc-4722-85d1-8cd196e055b3\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " Jan 26 17:26:15 crc kubenswrapper[4856]: I0126 17:26:15.249259 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vb88j\" (UniqueName: \"kubernetes.io/projected/77516bf5-33dc-4722-85d1-8cd196e055b3-kube-api-access-vb88j\") pod \"77516bf5-33dc-4722-85d1-8cd196e055b3\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " Jan 26 17:26:15 crc kubenswrapper[4856]: I0126 17:26:15.249280 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/77516bf5-33dc-4722-85d1-8cd196e055b3-buildcachedir\") pod \"77516bf5-33dc-4722-85d1-8cd196e055b3\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " Jan 26 17:26:15 crc kubenswrapper[4856]: I0126 17:26:15.249305 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/77516bf5-33dc-4722-85d1-8cd196e055b3-container-storage-root\") pod \"77516bf5-33dc-4722-85d1-8cd196e055b3\" (UID: \"77516bf5-33dc-4722-85d1-8cd196e055b3\") " Jan 26 17:26:15 crc kubenswrapper[4856]: I0126 17:26:15.249516 4856 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/77516bf5-33dc-4722-85d1-8cd196e055b3-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Jan 26 17:26:15 crc kubenswrapper[4856]: I0126 17:26:15.250480 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77516bf5-33dc-4722-85d1-8cd196e055b3-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "77516bf5-33dc-4722-85d1-8cd196e055b3" (UID: "77516bf5-33dc-4722-85d1-8cd196e055b3"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:26:15 crc kubenswrapper[4856]: I0126 17:26:15.250556 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/77516bf5-33dc-4722-85d1-8cd196e055b3-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "77516bf5-33dc-4722-85d1-8cd196e055b3" (UID: "77516bf5-33dc-4722-85d1-8cd196e055b3"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 17:26:15 crc kubenswrapper[4856]: I0126 17:26:15.251364 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77516bf5-33dc-4722-85d1-8cd196e055b3-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "77516bf5-33dc-4722-85d1-8cd196e055b3" (UID: "77516bf5-33dc-4722-85d1-8cd196e055b3"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:26:15 crc kubenswrapper[4856]: I0126 17:26:15.251973 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77516bf5-33dc-4722-85d1-8cd196e055b3-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "77516bf5-33dc-4722-85d1-8cd196e055b3" (UID: "77516bf5-33dc-4722-85d1-8cd196e055b3"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:26:15 crc kubenswrapper[4856]: I0126 17:26:15.253929 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77516bf5-33dc-4722-85d1-8cd196e055b3-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "77516bf5-33dc-4722-85d1-8cd196e055b3" (UID: "77516bf5-33dc-4722-85d1-8cd196e055b3"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:26:15 crc kubenswrapper[4856]: I0126 17:26:15.254169 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77516bf5-33dc-4722-85d1-8cd196e055b3-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "77516bf5-33dc-4722-85d1-8cd196e055b3" (UID: "77516bf5-33dc-4722-85d1-8cd196e055b3"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:26:15 crc kubenswrapper[4856]: I0126 17:26:15.256737 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77516bf5-33dc-4722-85d1-8cd196e055b3-kube-api-access-vb88j" (OuterVolumeSpecName: "kube-api-access-vb88j") pod "77516bf5-33dc-4722-85d1-8cd196e055b3" (UID: "77516bf5-33dc-4722-85d1-8cd196e055b3"). InnerVolumeSpecName "kube-api-access-vb88j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:26:15 crc kubenswrapper[4856]: I0126 17:26:15.257168 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77516bf5-33dc-4722-85d1-8cd196e055b3-builder-dockercfg-8h4xs-pull" (OuterVolumeSpecName: "builder-dockercfg-8h4xs-pull") pod "77516bf5-33dc-4722-85d1-8cd196e055b3" (UID: "77516bf5-33dc-4722-85d1-8cd196e055b3"). InnerVolumeSpecName "builder-dockercfg-8h4xs-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:26:15 crc kubenswrapper[4856]: I0126 17:26:15.257676 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77516bf5-33dc-4722-85d1-8cd196e055b3-builder-dockercfg-8h4xs-push" (OuterVolumeSpecName: "builder-dockercfg-8h4xs-push") pod "77516bf5-33dc-4722-85d1-8cd196e055b3" (UID: "77516bf5-33dc-4722-85d1-8cd196e055b3"). InnerVolumeSpecName "builder-dockercfg-8h4xs-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:26:15 crc kubenswrapper[4856]: I0126 17:26:15.350826 4856 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/77516bf5-33dc-4722-85d1-8cd196e055b3-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 26 17:26:15 crc kubenswrapper[4856]: I0126 17:26:15.350864 4856 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/77516bf5-33dc-4722-85d1-8cd196e055b3-buildworkdir\") on node \"crc\" DevicePath \"\"" Jan 26 17:26:15 crc kubenswrapper[4856]: I0126 17:26:15.350874 4856 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/77516bf5-33dc-4722-85d1-8cd196e055b3-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 26 17:26:15 crc kubenswrapper[4856]: I0126 17:26:15.350886 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vb88j\" (UniqueName: \"kubernetes.io/projected/77516bf5-33dc-4722-85d1-8cd196e055b3-kube-api-access-vb88j\") on node \"crc\" DevicePath \"\"" Jan 26 17:26:15 crc kubenswrapper[4856]: I0126 17:26:15.350897 4856 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/77516bf5-33dc-4722-85d1-8cd196e055b3-buildcachedir\") on node \"crc\" DevicePath \"\"" Jan 26 17:26:15 crc kubenswrapper[4856]: I0126 17:26:15.350907 4856 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-8h4xs-pull\" (UniqueName: \"kubernetes.io/secret/77516bf5-33dc-4722-85d1-8cd196e055b3-builder-dockercfg-8h4xs-pull\") on node \"crc\" DevicePath \"\"" Jan 26 17:26:15 crc kubenswrapper[4856]: I0126 17:26:15.350917 4856 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-8h4xs-push\" (UniqueName: \"kubernetes.io/secret/77516bf5-33dc-4722-85d1-8cd196e055b3-builder-dockercfg-8h4xs-push\") on node \"crc\" DevicePath \"\"" Jan 26 17:26:15 crc kubenswrapper[4856]: I0126 17:26:15.351126 4856 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/77516bf5-33dc-4722-85d1-8cd196e055b3-build-system-configs\") on node \"crc\" DevicePath \"\"" Jan 26 17:26:15 crc kubenswrapper[4856]: I0126 17:26:15.351139 4856 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/77516bf5-33dc-4722-85d1-8cd196e055b3-container-storage-run\") on node \"crc\" DevicePath \"\"" Jan 26 17:26:15 crc kubenswrapper[4856]: I0126 17:26:15.353121 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77516bf5-33dc-4722-85d1-8cd196e055b3-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "77516bf5-33dc-4722-85d1-8cd196e055b3" (UID: "77516bf5-33dc-4722-85d1-8cd196e055b3"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:26:15 crc kubenswrapper[4856]: I0126 17:26:15.452776 4856 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/77516bf5-33dc-4722-85d1-8cd196e055b3-build-blob-cache\") on node \"crc\" DevicePath \"\"" Jan 26 17:26:15 crc kubenswrapper[4856]: I0126 17:26:15.822432 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77516bf5-33dc-4722-85d1-8cd196e055b3-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "77516bf5-33dc-4722-85d1-8cd196e055b3" (UID: "77516bf5-33dc-4722-85d1-8cd196e055b3"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:26:15 crc kubenswrapper[4856]: I0126 17:26:15.857412 4856 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/77516bf5-33dc-4722-85d1-8cd196e055b3-container-storage-root\") on node \"crc\" DevicePath \"\"" Jan 26 17:26:15 crc kubenswrapper[4856]: I0126 17:26:15.885062 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-webhook-snmp-5-build_77516bf5-33dc-4722-85d1-8cd196e055b3/docker-build/0.log" Jan 26 17:26:15 crc kubenswrapper[4856]: I0126 17:26:15.885908 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-5-build" event={"ID":"77516bf5-33dc-4722-85d1-8cd196e055b3","Type":"ContainerDied","Data":"73566cebe45397c2722640fc7ffd2e04eb8a444fc2c63f9f96879b8d7aec440b"} Jan 26 17:26:15 crc kubenswrapper[4856]: I0126 17:26:15.885941 4856 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73566cebe45397c2722640fc7ffd2e04eb8a444fc2c63f9f96879b8d7aec440b" Jan 26 17:26:15 crc kubenswrapper[4856]: I0126 17:26:15.885980 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-5-build" Jan 26 17:26:26 crc kubenswrapper[4856]: I0126 17:26:26.939507 4856 patch_prober.go:28] interesting pod/machine-config-daemon-xm9cq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:26:26 crc kubenswrapper[4856]: I0126 17:26:26.940155 4856 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:26:29 crc kubenswrapper[4856]: I0126 17:26:29.671606 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hwx7g"] Jan 26 17:26:29 crc kubenswrapper[4856]: E0126 17:26:29.672229 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77516bf5-33dc-4722-85d1-8cd196e055b3" containerName="manage-dockerfile" Jan 26 17:26:29 crc kubenswrapper[4856]: I0126 17:26:29.672246 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="77516bf5-33dc-4722-85d1-8cd196e055b3" containerName="manage-dockerfile" Jan 26 17:26:29 crc kubenswrapper[4856]: E0126 17:26:29.672264 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77516bf5-33dc-4722-85d1-8cd196e055b3" containerName="git-clone" Jan 26 17:26:29 crc kubenswrapper[4856]: I0126 17:26:29.672273 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="77516bf5-33dc-4722-85d1-8cd196e055b3" containerName="git-clone" Jan 26 17:26:29 crc kubenswrapper[4856]: E0126 17:26:29.672295 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77516bf5-33dc-4722-85d1-8cd196e055b3" containerName="docker-build" Jan 26 17:26:29 crc kubenswrapper[4856]: I0126 17:26:29.672302 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="77516bf5-33dc-4722-85d1-8cd196e055b3" containerName="docker-build" Jan 26 17:26:29 crc kubenswrapper[4856]: I0126 17:26:29.672403 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="77516bf5-33dc-4722-85d1-8cd196e055b3" containerName="docker-build" Jan 26 17:26:29 crc kubenswrapper[4856]: I0126 17:26:29.674317 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hwx7g" Jan 26 17:26:29 crc kubenswrapper[4856]: I0126 17:26:29.679261 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hwx7g"] Jan 26 17:26:29 crc kubenswrapper[4856]: I0126 17:26:29.857500 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44rwk\" (UniqueName: \"kubernetes.io/projected/a26e6806-50dc-4680-a6d4-01f73fce49da-kube-api-access-44rwk\") pod \"certified-operators-hwx7g\" (UID: \"a26e6806-50dc-4680-a6d4-01f73fce49da\") " pod="openshift-marketplace/certified-operators-hwx7g" Jan 26 17:26:29 crc kubenswrapper[4856]: I0126 17:26:29.857578 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a26e6806-50dc-4680-a6d4-01f73fce49da-catalog-content\") pod \"certified-operators-hwx7g\" (UID: \"a26e6806-50dc-4680-a6d4-01f73fce49da\") " pod="openshift-marketplace/certified-operators-hwx7g" Jan 26 17:26:29 crc kubenswrapper[4856]: I0126 17:26:29.857612 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a26e6806-50dc-4680-a6d4-01f73fce49da-utilities\") pod \"certified-operators-hwx7g\" (UID: \"a26e6806-50dc-4680-a6d4-01f73fce49da\") " pod="openshift-marketplace/certified-operators-hwx7g" Jan 26 17:26:29 crc kubenswrapper[4856]: I0126 17:26:29.959389 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44rwk\" (UniqueName: \"kubernetes.io/projected/a26e6806-50dc-4680-a6d4-01f73fce49da-kube-api-access-44rwk\") pod \"certified-operators-hwx7g\" (UID: \"a26e6806-50dc-4680-a6d4-01f73fce49da\") " pod="openshift-marketplace/certified-operators-hwx7g" Jan 26 17:26:29 crc kubenswrapper[4856]: I0126 17:26:29.959444 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a26e6806-50dc-4680-a6d4-01f73fce49da-catalog-content\") pod \"certified-operators-hwx7g\" (UID: \"a26e6806-50dc-4680-a6d4-01f73fce49da\") " pod="openshift-marketplace/certified-operators-hwx7g" Jan 26 17:26:29 crc kubenswrapper[4856]: I0126 17:26:29.959468 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a26e6806-50dc-4680-a6d4-01f73fce49da-utilities\") pod \"certified-operators-hwx7g\" (UID: \"a26e6806-50dc-4680-a6d4-01f73fce49da\") " pod="openshift-marketplace/certified-operators-hwx7g" Jan 26 17:26:29 crc kubenswrapper[4856]: I0126 17:26:29.960067 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a26e6806-50dc-4680-a6d4-01f73fce49da-catalog-content\") pod \"certified-operators-hwx7g\" (UID: \"a26e6806-50dc-4680-a6d4-01f73fce49da\") " pod="openshift-marketplace/certified-operators-hwx7g" Jan 26 17:26:29 crc kubenswrapper[4856]: I0126 17:26:29.960236 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a26e6806-50dc-4680-a6d4-01f73fce49da-utilities\") pod \"certified-operators-hwx7g\" (UID: \"a26e6806-50dc-4680-a6d4-01f73fce49da\") " pod="openshift-marketplace/certified-operators-hwx7g" Jan 26 17:26:29 crc kubenswrapper[4856]: I0126 17:26:29.989510 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44rwk\" (UniqueName: \"kubernetes.io/projected/a26e6806-50dc-4680-a6d4-01f73fce49da-kube-api-access-44rwk\") pod \"certified-operators-hwx7g\" (UID: \"a26e6806-50dc-4680-a6d4-01f73fce49da\") " pod="openshift-marketplace/certified-operators-hwx7g" Jan 26 17:26:29 crc kubenswrapper[4856]: I0126 17:26:29.992070 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hwx7g" Jan 26 17:26:30 crc kubenswrapper[4856]: I0126 17:26:30.478559 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hwx7g"] Jan 26 17:26:31 crc kubenswrapper[4856]: I0126 17:26:31.004661 4856 generic.go:334] "Generic (PLEG): container finished" podID="a26e6806-50dc-4680-a6d4-01f73fce49da" containerID="416cabdd5226bae8c40dffba6482cf11182130853f71df6b638445e7408ef751" exitCode=0 Jan 26 17:26:31 crc kubenswrapper[4856]: I0126 17:26:31.004726 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hwx7g" event={"ID":"a26e6806-50dc-4680-a6d4-01f73fce49da","Type":"ContainerDied","Data":"416cabdd5226bae8c40dffba6482cf11182130853f71df6b638445e7408ef751"} Jan 26 17:26:31 crc kubenswrapper[4856]: I0126 17:26:31.005102 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hwx7g" event={"ID":"a26e6806-50dc-4680-a6d4-01f73fce49da","Type":"ContainerStarted","Data":"ecdf439a6fbc2c2d5caa5779f36a413aa1c7d30d9a0374701a3ccdd3bf9dacd1"} Jan 26 17:26:35 crc kubenswrapper[4856]: I0126 17:26:35.046741 4856 generic.go:334] "Generic (PLEG): container finished" podID="a26e6806-50dc-4680-a6d4-01f73fce49da" containerID="5f00836821d46247c6984dd8810c92bcfea2ef9f97d77ed0de71b6b21ef7919c" exitCode=0 Jan 26 17:26:35 crc kubenswrapper[4856]: I0126 17:26:35.046823 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hwx7g" event={"ID":"a26e6806-50dc-4680-a6d4-01f73fce49da","Type":"ContainerDied","Data":"5f00836821d46247c6984dd8810c92bcfea2ef9f97d77ed0de71b6b21ef7919c"} Jan 26 17:26:36 crc kubenswrapper[4856]: I0126 17:26:36.057050 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hwx7g" event={"ID":"a26e6806-50dc-4680-a6d4-01f73fce49da","Type":"ContainerStarted","Data":"9d274f3313f18a3beb64c0ccabe70ee5b097ee7ee5ee0b5b90ae0dc6885d8137"} Jan 26 17:26:36 crc kubenswrapper[4856]: I0126 17:26:36.075520 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hwx7g" podStartSLOduration=2.5638584079999998 podStartE2EDuration="7.075486858s" podCreationTimestamp="2026-01-26 17:26:29 +0000 UTC" firstStartedPulling="2026-01-26 17:26:31.006423583 +0000 UTC m=+1686.959677564" lastFinishedPulling="2026-01-26 17:26:35.518052033 +0000 UTC m=+1691.471306014" observedRunningTime="2026-01-26 17:26:36.072181613 +0000 UTC m=+1692.025435634" watchObservedRunningTime="2026-01-26 17:26:36.075486858 +0000 UTC m=+1692.028740879" Jan 26 17:26:39 crc kubenswrapper[4856]: I0126 17:26:39.993178 4856 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hwx7g" Jan 26 17:26:39 crc kubenswrapper[4856]: I0126 17:26:39.994181 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hwx7g" Jan 26 17:26:40 crc kubenswrapper[4856]: I0126 17:26:40.060713 4856 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hwx7g" Jan 26 17:26:40 crc kubenswrapper[4856]: I0126 17:26:40.142993 4856 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hwx7g" Jan 26 17:26:40 crc kubenswrapper[4856]: I0126 17:26:40.302583 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hwx7g"] Jan 26 17:26:40 crc kubenswrapper[4856]: I0126 17:26:40.362035 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bxhpt"] Jan 26 17:26:40 crc kubenswrapper[4856]: I0126 17:26:40.362354 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bxhpt" podUID="5f34c6a8-6023-480c-a25e-46f8c4f3766b" containerName="registry-server" containerID="cri-o://094471a22cb2be3806f3e0c5d6e07e36d6a5a423b089fe73b28707ddde1dde10" gracePeriod=2 Jan 26 17:26:41 crc kubenswrapper[4856]: I0126 17:26:41.908862 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bxhpt" Jan 26 17:26:42 crc kubenswrapper[4856]: I0126 17:26:42.031357 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f34c6a8-6023-480c-a25e-46f8c4f3766b-utilities\") pod \"5f34c6a8-6023-480c-a25e-46f8c4f3766b\" (UID: \"5f34c6a8-6023-480c-a25e-46f8c4f3766b\") " Jan 26 17:26:42 crc kubenswrapper[4856]: I0126 17:26:42.031486 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f34c6a8-6023-480c-a25e-46f8c4f3766b-catalog-content\") pod \"5f34c6a8-6023-480c-a25e-46f8c4f3766b\" (UID: \"5f34c6a8-6023-480c-a25e-46f8c4f3766b\") " Jan 26 17:26:42 crc kubenswrapper[4856]: I0126 17:26:42.031536 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crdtn\" (UniqueName: \"kubernetes.io/projected/5f34c6a8-6023-480c-a25e-46f8c4f3766b-kube-api-access-crdtn\") pod \"5f34c6a8-6023-480c-a25e-46f8c4f3766b\" (UID: \"5f34c6a8-6023-480c-a25e-46f8c4f3766b\") " Jan 26 17:26:42 crc kubenswrapper[4856]: I0126 17:26:42.032082 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f34c6a8-6023-480c-a25e-46f8c4f3766b-utilities" (OuterVolumeSpecName: "utilities") pod "5f34c6a8-6023-480c-a25e-46f8c4f3766b" (UID: "5f34c6a8-6023-480c-a25e-46f8c4f3766b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:26:42 crc kubenswrapper[4856]: I0126 17:26:42.037340 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f34c6a8-6023-480c-a25e-46f8c4f3766b-kube-api-access-crdtn" (OuterVolumeSpecName: "kube-api-access-crdtn") pod "5f34c6a8-6023-480c-a25e-46f8c4f3766b" (UID: "5f34c6a8-6023-480c-a25e-46f8c4f3766b"). InnerVolumeSpecName "kube-api-access-crdtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:26:42 crc kubenswrapper[4856]: I0126 17:26:42.075665 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f34c6a8-6023-480c-a25e-46f8c4f3766b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5f34c6a8-6023-480c-a25e-46f8c4f3766b" (UID: "5f34c6a8-6023-480c-a25e-46f8c4f3766b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:26:42 crc kubenswrapper[4856]: I0126 17:26:42.104354 4856 generic.go:334] "Generic (PLEG): container finished" podID="5f34c6a8-6023-480c-a25e-46f8c4f3766b" containerID="094471a22cb2be3806f3e0c5d6e07e36d6a5a423b089fe73b28707ddde1dde10" exitCode=0 Jan 26 17:26:42 crc kubenswrapper[4856]: I0126 17:26:42.104758 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bxhpt" event={"ID":"5f34c6a8-6023-480c-a25e-46f8c4f3766b","Type":"ContainerDied","Data":"094471a22cb2be3806f3e0c5d6e07e36d6a5a423b089fe73b28707ddde1dde10"} Jan 26 17:26:42 crc kubenswrapper[4856]: I0126 17:26:42.104842 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bxhpt" Jan 26 17:26:42 crc kubenswrapper[4856]: I0126 17:26:42.104873 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bxhpt" event={"ID":"5f34c6a8-6023-480c-a25e-46f8c4f3766b","Type":"ContainerDied","Data":"ae868d389f2d56b098915bbed54fc03534f4fd1519a0d344eda69f6356db31f0"} Jan 26 17:26:42 crc kubenswrapper[4856]: I0126 17:26:42.104982 4856 scope.go:117] "RemoveContainer" containerID="094471a22cb2be3806f3e0c5d6e07e36d6a5a423b089fe73b28707ddde1dde10" Jan 26 17:26:42 crc kubenswrapper[4856]: I0126 17:26:42.141860 4856 scope.go:117] "RemoveContainer" containerID="ab4dc2168cf3030f71b121144c073ab78a14a965d8feaf1a0de933f786e1cb89" Jan 26 17:26:42 crc kubenswrapper[4856]: I0126 17:26:42.143855 4856 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f34c6a8-6023-480c-a25e-46f8c4f3766b-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 17:26:42 crc kubenswrapper[4856]: I0126 17:26:42.143902 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crdtn\" (UniqueName: \"kubernetes.io/projected/5f34c6a8-6023-480c-a25e-46f8c4f3766b-kube-api-access-crdtn\") on node \"crc\" DevicePath \"\"" Jan 26 17:26:42 crc kubenswrapper[4856]: I0126 17:26:42.143915 4856 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f34c6a8-6023-480c-a25e-46f8c4f3766b-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 17:26:42 crc kubenswrapper[4856]: I0126 17:26:42.147687 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bxhpt"] Jan 26 17:26:42 crc kubenswrapper[4856]: I0126 17:26:42.152358 4856 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bxhpt"] Jan 26 17:26:42 crc kubenswrapper[4856]: I0126 17:26:42.175498 4856 scope.go:117] "RemoveContainer" containerID="405a5e16f820190605a6762d0b7653fa1a6bedd12b761afcd55093867a05ee57" Jan 26 17:26:42 crc kubenswrapper[4856]: I0126 17:26:42.197875 4856 scope.go:117] "RemoveContainer" containerID="094471a22cb2be3806f3e0c5d6e07e36d6a5a423b089fe73b28707ddde1dde10" Jan 26 17:26:42 crc kubenswrapper[4856]: E0126 17:26:42.198455 4856 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"094471a22cb2be3806f3e0c5d6e07e36d6a5a423b089fe73b28707ddde1dde10\": container with ID starting with 094471a22cb2be3806f3e0c5d6e07e36d6a5a423b089fe73b28707ddde1dde10 not found: ID does not exist" containerID="094471a22cb2be3806f3e0c5d6e07e36d6a5a423b089fe73b28707ddde1dde10" Jan 26 17:26:42 crc kubenswrapper[4856]: I0126 17:26:42.198610 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"094471a22cb2be3806f3e0c5d6e07e36d6a5a423b089fe73b28707ddde1dde10"} err="failed to get container status \"094471a22cb2be3806f3e0c5d6e07e36d6a5a423b089fe73b28707ddde1dde10\": rpc error: code = NotFound desc = could not find container \"094471a22cb2be3806f3e0c5d6e07e36d6a5a423b089fe73b28707ddde1dde10\": container with ID starting with 094471a22cb2be3806f3e0c5d6e07e36d6a5a423b089fe73b28707ddde1dde10 not found: ID does not exist" Jan 26 17:26:42 crc kubenswrapper[4856]: I0126 17:26:42.198713 4856 scope.go:117] "RemoveContainer" containerID="ab4dc2168cf3030f71b121144c073ab78a14a965d8feaf1a0de933f786e1cb89" Jan 26 17:26:42 crc kubenswrapper[4856]: E0126 17:26:42.199256 4856 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab4dc2168cf3030f71b121144c073ab78a14a965d8feaf1a0de933f786e1cb89\": container with ID starting with ab4dc2168cf3030f71b121144c073ab78a14a965d8feaf1a0de933f786e1cb89 not found: ID does not exist" containerID="ab4dc2168cf3030f71b121144c073ab78a14a965d8feaf1a0de933f786e1cb89" Jan 26 17:26:42 crc kubenswrapper[4856]: I0126 17:26:42.199353 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab4dc2168cf3030f71b121144c073ab78a14a965d8feaf1a0de933f786e1cb89"} err="failed to get container status \"ab4dc2168cf3030f71b121144c073ab78a14a965d8feaf1a0de933f786e1cb89\": rpc error: code = NotFound desc = could not find container \"ab4dc2168cf3030f71b121144c073ab78a14a965d8feaf1a0de933f786e1cb89\": container with ID starting with ab4dc2168cf3030f71b121144c073ab78a14a965d8feaf1a0de933f786e1cb89 not found: ID does not exist" Jan 26 17:26:42 crc kubenswrapper[4856]: I0126 17:26:42.199461 4856 scope.go:117] "RemoveContainer" containerID="405a5e16f820190605a6762d0b7653fa1a6bedd12b761afcd55093867a05ee57" Jan 26 17:26:42 crc kubenswrapper[4856]: E0126 17:26:42.199853 4856 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"405a5e16f820190605a6762d0b7653fa1a6bedd12b761afcd55093867a05ee57\": container with ID starting with 405a5e16f820190605a6762d0b7653fa1a6bedd12b761afcd55093867a05ee57 not found: ID does not exist" containerID="405a5e16f820190605a6762d0b7653fa1a6bedd12b761afcd55093867a05ee57" Jan 26 17:26:42 crc kubenswrapper[4856]: I0126 17:26:42.199903 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"405a5e16f820190605a6762d0b7653fa1a6bedd12b761afcd55093867a05ee57"} err="failed to get container status \"405a5e16f820190605a6762d0b7653fa1a6bedd12b761afcd55093867a05ee57\": rpc error: code = NotFound desc = could not find container \"405a5e16f820190605a6762d0b7653fa1a6bedd12b761afcd55093867a05ee57\": container with ID starting with 405a5e16f820190605a6762d0b7653fa1a6bedd12b761afcd55093867a05ee57 not found: ID does not exist" Jan 26 17:26:43 crc kubenswrapper[4856]: I0126 17:26:43.403965 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f34c6a8-6023-480c-a25e-46f8c4f3766b" path="/var/lib/kubelet/pods/5f34c6a8-6023-480c-a25e-46f8c4f3766b/volumes" Jan 26 17:26:56 crc kubenswrapper[4856]: I0126 17:26:56.938996 4856 patch_prober.go:28] interesting pod/machine-config-daemon-xm9cq container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 17:26:56 crc kubenswrapper[4856]: I0126 17:26:56.939787 4856 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 17:26:56 crc kubenswrapper[4856]: I0126 17:26:56.939854 4856 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" Jan 26 17:26:56 crc kubenswrapper[4856]: I0126 17:26:56.941050 4856 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b8175a0e79754a858867d9a98f2aa6c52214536db6005b6724cf907eb7a891ee"} pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 26 17:26:56 crc kubenswrapper[4856]: I0126 17:26:56.941219 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" containerName="machine-config-daemon" containerID="cri-o://b8175a0e79754a858867d9a98f2aa6c52214536db6005b6724cf907eb7a891ee" gracePeriod=600 Jan 26 17:26:59 crc kubenswrapper[4856]: E0126 17:26:59.854476 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xm9cq_openshift-machine-config-operator(63c75ede-5170-4db0-811b-5217ef8d72b3)\"" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" Jan 26 17:27:00 crc kubenswrapper[4856]: I0126 17:27:00.250219 4856 generic.go:334] "Generic (PLEG): container finished" podID="63c75ede-5170-4db0-811b-5217ef8d72b3" containerID="b8175a0e79754a858867d9a98f2aa6c52214536db6005b6724cf907eb7a891ee" exitCode=0 Jan 26 17:27:00 crc kubenswrapper[4856]: I0126 17:27:00.250299 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" event={"ID":"63c75ede-5170-4db0-811b-5217ef8d72b3","Type":"ContainerDied","Data":"b8175a0e79754a858867d9a98f2aa6c52214536db6005b6724cf907eb7a891ee"} Jan 26 17:27:00 crc kubenswrapper[4856]: I0126 17:27:00.250475 4856 scope.go:117] "RemoveContainer" containerID="cda3cdbac0b1e3c460ee9a5617b9c5fd59d4db5c67a69b81c9224934be12563c" Jan 26 17:27:00 crc kubenswrapper[4856]: I0126 17:27:00.251031 4856 scope.go:117] "RemoveContainer" containerID="b8175a0e79754a858867d9a98f2aa6c52214536db6005b6724cf907eb7a891ee" Jan 26 17:27:00 crc kubenswrapper[4856]: E0126 17:27:00.252746 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xm9cq_openshift-machine-config-operator(63c75ede-5170-4db0-811b-5217ef8d72b3)\"" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" Jan 26 17:27:06 crc kubenswrapper[4856]: I0126 17:27:06.134582 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-pxr8g/must-gather-xhb5b"] Jan 26 17:27:06 crc kubenswrapper[4856]: E0126 17:27:06.135599 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f34c6a8-6023-480c-a25e-46f8c4f3766b" containerName="extract-utilities" Jan 26 17:27:06 crc kubenswrapper[4856]: I0126 17:27:06.135630 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f34c6a8-6023-480c-a25e-46f8c4f3766b" containerName="extract-utilities" Jan 26 17:27:06 crc kubenswrapper[4856]: E0126 17:27:06.135653 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f34c6a8-6023-480c-a25e-46f8c4f3766b" containerName="registry-server" Jan 26 17:27:06 crc kubenswrapper[4856]: I0126 17:27:06.135662 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f34c6a8-6023-480c-a25e-46f8c4f3766b" containerName="registry-server" Jan 26 17:27:06 crc kubenswrapper[4856]: E0126 17:27:06.135678 4856 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f34c6a8-6023-480c-a25e-46f8c4f3766b" containerName="extract-content" Jan 26 17:27:06 crc kubenswrapper[4856]: I0126 17:27:06.135688 4856 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f34c6a8-6023-480c-a25e-46f8c4f3766b" containerName="extract-content" Jan 26 17:27:06 crc kubenswrapper[4856]: I0126 17:27:06.135913 4856 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f34c6a8-6023-480c-a25e-46f8c4f3766b" containerName="registry-server" Jan 26 17:27:06 crc kubenswrapper[4856]: I0126 17:27:06.136909 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pxr8g/must-gather-xhb5b" Jan 26 17:27:06 crc kubenswrapper[4856]: I0126 17:27:06.139385 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-pxr8g"/"kube-root-ca.crt" Jan 26 17:27:06 crc kubenswrapper[4856]: I0126 17:27:06.139782 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-pxr8g"/"default-dockercfg-jpqwb" Jan 26 17:27:06 crc kubenswrapper[4856]: I0126 17:27:06.142463 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-pxr8g"/"openshift-service-ca.crt" Jan 26 17:27:06 crc kubenswrapper[4856]: I0126 17:27:06.202948 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-pxr8g/must-gather-xhb5b"] Jan 26 17:27:06 crc kubenswrapper[4856]: I0126 17:27:06.243139 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0190f6f3-2762-486f-8648-e59c9e5179ad-must-gather-output\") pod \"must-gather-xhb5b\" (UID: \"0190f6f3-2762-486f-8648-e59c9e5179ad\") " pod="openshift-must-gather-pxr8g/must-gather-xhb5b" Jan 26 17:27:06 crc kubenswrapper[4856]: I0126 17:27:06.243194 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vms6d\" (UniqueName: \"kubernetes.io/projected/0190f6f3-2762-486f-8648-e59c9e5179ad-kube-api-access-vms6d\") pod \"must-gather-xhb5b\" (UID: \"0190f6f3-2762-486f-8648-e59c9e5179ad\") " pod="openshift-must-gather-pxr8g/must-gather-xhb5b" Jan 26 17:27:06 crc kubenswrapper[4856]: I0126 17:27:06.344151 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vms6d\" (UniqueName: \"kubernetes.io/projected/0190f6f3-2762-486f-8648-e59c9e5179ad-kube-api-access-vms6d\") pod \"must-gather-xhb5b\" (UID: \"0190f6f3-2762-486f-8648-e59c9e5179ad\") " pod="openshift-must-gather-pxr8g/must-gather-xhb5b" Jan 26 17:27:06 crc kubenswrapper[4856]: I0126 17:27:06.344255 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0190f6f3-2762-486f-8648-e59c9e5179ad-must-gather-output\") pod \"must-gather-xhb5b\" (UID: \"0190f6f3-2762-486f-8648-e59c9e5179ad\") " pod="openshift-must-gather-pxr8g/must-gather-xhb5b" Jan 26 17:27:06 crc kubenswrapper[4856]: I0126 17:27:06.344689 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0190f6f3-2762-486f-8648-e59c9e5179ad-must-gather-output\") pod \"must-gather-xhb5b\" (UID: \"0190f6f3-2762-486f-8648-e59c9e5179ad\") " pod="openshift-must-gather-pxr8g/must-gather-xhb5b" Jan 26 17:27:06 crc kubenswrapper[4856]: I0126 17:27:06.371410 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vms6d\" (UniqueName: \"kubernetes.io/projected/0190f6f3-2762-486f-8648-e59c9e5179ad-kube-api-access-vms6d\") pod \"must-gather-xhb5b\" (UID: \"0190f6f3-2762-486f-8648-e59c9e5179ad\") " pod="openshift-must-gather-pxr8g/must-gather-xhb5b" Jan 26 17:27:06 crc kubenswrapper[4856]: I0126 17:27:06.529804 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pxr8g/must-gather-xhb5b" Jan 26 17:27:06 crc kubenswrapper[4856]: I0126 17:27:06.805813 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-pxr8g/must-gather-xhb5b"] Jan 26 17:27:07 crc kubenswrapper[4856]: I0126 17:27:07.304889 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pxr8g/must-gather-xhb5b" event={"ID":"0190f6f3-2762-486f-8648-e59c9e5179ad","Type":"ContainerStarted","Data":"3ce26f76fc446d8a550d3746649cb18a49ea0c18d56a871d55ff91fd1f0a4363"} Jan 26 17:27:12 crc kubenswrapper[4856]: I0126 17:27:12.395456 4856 scope.go:117] "RemoveContainer" containerID="b8175a0e79754a858867d9a98f2aa6c52214536db6005b6724cf907eb7a891ee" Jan 26 17:27:12 crc kubenswrapper[4856]: E0126 17:27:12.396112 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xm9cq_openshift-machine-config-operator(63c75ede-5170-4db0-811b-5217ef8d72b3)\"" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" Jan 26 17:27:20 crc kubenswrapper[4856]: I0126 17:27:20.449287 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pxr8g/must-gather-xhb5b" event={"ID":"0190f6f3-2762-486f-8648-e59c9e5179ad","Type":"ContainerStarted","Data":"3f464ae4156c51b0e76930516abce0275686a5ccb1e2b35e0204d8a9e05e513a"} Jan 26 17:27:20 crc kubenswrapper[4856]: I0126 17:27:20.449873 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pxr8g/must-gather-xhb5b" event={"ID":"0190f6f3-2762-486f-8648-e59c9e5179ad","Type":"ContainerStarted","Data":"1c6e1c2b24da05f9138f08890b3f71dfa17ade72ec99276a5860cbc9c8b26e99"} Jan 26 17:27:20 crc kubenswrapper[4856]: I0126 17:27:20.467647 4856 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-pxr8g/must-gather-xhb5b" podStartSLOduration=1.827602512 podStartE2EDuration="14.467609024s" podCreationTimestamp="2026-01-26 17:27:06 +0000 UTC" firstStartedPulling="2026-01-26 17:27:06.811204187 +0000 UTC m=+1722.764458168" lastFinishedPulling="2026-01-26 17:27:19.451210699 +0000 UTC m=+1735.404464680" observedRunningTime="2026-01-26 17:27:20.46221228 +0000 UTC m=+1736.415466291" watchObservedRunningTime="2026-01-26 17:27:20.467609024 +0000 UTC m=+1736.420863015" Jan 26 17:27:25 crc kubenswrapper[4856]: I0126 17:27:25.400650 4856 scope.go:117] "RemoveContainer" containerID="b8175a0e79754a858867d9a98f2aa6c52214536db6005b6724cf907eb7a891ee" Jan 26 17:27:25 crc kubenswrapper[4856]: E0126 17:27:25.401561 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xm9cq_openshift-machine-config-operator(63c75ede-5170-4db0-811b-5217ef8d72b3)\"" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" Jan 26 17:27:40 crc kubenswrapper[4856]: I0126 17:27:40.395739 4856 scope.go:117] "RemoveContainer" containerID="b8175a0e79754a858867d9a98f2aa6c52214536db6005b6724cf907eb7a891ee" Jan 26 17:27:40 crc kubenswrapper[4856]: E0126 17:27:40.396381 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xm9cq_openshift-machine-config-operator(63c75ede-5170-4db0-811b-5217ef8d72b3)\"" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" Jan 26 17:27:43 crc kubenswrapper[4856]: I0126 17:27:43.868093 4856 scope.go:117] "RemoveContainer" containerID="1f422f325f8fa610d27451f9734bb53bd3fdded1c3e711a971293ebcfe442247" Jan 26 17:27:43 crc kubenswrapper[4856]: I0126 17:27:43.894439 4856 scope.go:117] "RemoveContainer" containerID="37c0850dfb46215185250da7eccd6ad1561ff7e374d7cfc24a1386d6bf8bcf2f" Jan 26 17:27:52 crc kubenswrapper[4856]: I0126 17:27:52.396728 4856 scope.go:117] "RemoveContainer" containerID="b8175a0e79754a858867d9a98f2aa6c52214536db6005b6724cf907eb7a891ee" Jan 26 17:27:52 crc kubenswrapper[4856]: E0126 17:27:52.397457 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xm9cq_openshift-machine-config-operator(63c75ede-5170-4db0-811b-5217ef8d72b3)\"" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" Jan 26 17:28:03 crc kubenswrapper[4856]: I0126 17:28:03.395329 4856 scope.go:117] "RemoveContainer" containerID="b8175a0e79754a858867d9a98f2aa6c52214536db6005b6724cf907eb7a891ee" Jan 26 17:28:03 crc kubenswrapper[4856]: E0126 17:28:03.396138 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xm9cq_openshift-machine-config-operator(63c75ede-5170-4db0-811b-5217ef8d72b3)\"" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" Jan 26 17:28:03 crc kubenswrapper[4856]: I0126 17:28:03.771853 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-rrhjv_a1546392-4a69-4b12-8d7e-97450b73b7ca/control-plane-machine-set-operator/0.log" Jan 26 17:28:03 crc kubenswrapper[4856]: I0126 17:28:03.911930 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-7xb2b_77a97acb-2908-48fb-8bcd-0647f3e90160/kube-rbac-proxy/0.log" Jan 26 17:28:03 crc kubenswrapper[4856]: I0126 17:28:03.926453 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-7xb2b_77a97acb-2908-48fb-8bcd-0647f3e90160/machine-api-operator/0.log" Jan 26 17:28:14 crc kubenswrapper[4856]: I0126 17:28:14.394782 4856 scope.go:117] "RemoveContainer" containerID="b8175a0e79754a858867d9a98f2aa6c52214536db6005b6724cf907eb7a891ee" Jan 26 17:28:14 crc kubenswrapper[4856]: E0126 17:28:14.395461 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xm9cq_openshift-machine-config-operator(63c75ede-5170-4db0-811b-5217ef8d72b3)\"" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" Jan 26 17:28:17 crc kubenswrapper[4856]: I0126 17:28:17.586730 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-86cb77c54b-xcqr4_3dc10d6b-aa48-4c7d-afab-45fa62298819/cert-manager-controller/0.log" Jan 26 17:28:17 crc kubenswrapper[4856]: I0126 17:28:17.709096 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-855d9ccff4-rm9wd_24a9d780-2b57-49d2-9cb9-eac2456ed86d/cert-manager-cainjector/0.log" Jan 26 17:28:17 crc kubenswrapper[4856]: I0126 17:28:17.783922 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-f4fb5df64-www8b_e9288910-baf7-4cc4-b313-c87b80bfdd3e/cert-manager-webhook/0.log" Jan 26 17:28:28 crc kubenswrapper[4856]: I0126 17:28:28.395754 4856 scope.go:117] "RemoveContainer" containerID="b8175a0e79754a858867d9a98f2aa6c52214536db6005b6724cf907eb7a891ee" Jan 26 17:28:28 crc kubenswrapper[4856]: E0126 17:28:28.396566 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xm9cq_openshift-machine-config-operator(63c75ede-5170-4db0-811b-5217ef8d72b3)\"" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" Jan 26 17:28:31 crc kubenswrapper[4856]: I0126 17:28:31.270745 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-cq2gx_e31d2d53-8992-45e3-98aa-24ea73236248/prometheus-operator/0.log" Jan 26 17:28:31 crc kubenswrapper[4856]: I0126 17:28:31.516544 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-68d9fdc4dd-cf7wn_7c88687f-1304-4709-b148-a196f0d0190d/prometheus-operator-admission-webhook/0.log" Jan 26 17:28:31 crc kubenswrapper[4856]: I0126 17:28:31.544616 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-68d9fdc4dd-jbq25_766f50ba-0751-4f25-a6db-3b7195e72f55/prometheus-operator-admission-webhook/0.log" Jan 26 17:28:31 crc kubenswrapper[4856]: I0126 17:28:31.786856 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-fpn2h_a4ae7646-2afb-4ada-b8a4-d20a69f87949/operator/0.log" Jan 26 17:28:31 crc kubenswrapper[4856]: I0126 17:28:31.795543 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-5bmfp_bd7597f2-d44b-4e1b-ac60-b409985e3351/perses-operator/0.log" Jan 26 17:28:42 crc kubenswrapper[4856]: I0126 17:28:42.395851 4856 scope.go:117] "RemoveContainer" containerID="b8175a0e79754a858867d9a98f2aa6c52214536db6005b6724cf907eb7a891ee" Jan 26 17:28:42 crc kubenswrapper[4856]: E0126 17:28:42.396796 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xm9cq_openshift-machine-config-operator(63c75ede-5170-4db0-811b-5217ef8d72b3)\"" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" Jan 26 17:28:45 crc kubenswrapper[4856]: I0126 17:28:45.520258 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ank5l6_6521dc23-8f4e-452f-ae3e-167424fa3ed2/util/0.log" Jan 26 17:28:45 crc kubenswrapper[4856]: I0126 17:28:45.741401 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ank5l6_6521dc23-8f4e-452f-ae3e-167424fa3ed2/pull/0.log" Jan 26 17:28:45 crc kubenswrapper[4856]: I0126 17:28:45.760251 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ank5l6_6521dc23-8f4e-452f-ae3e-167424fa3ed2/util/0.log" Jan 26 17:28:45 crc kubenswrapper[4856]: I0126 17:28:45.766905 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ank5l6_6521dc23-8f4e-452f-ae3e-167424fa3ed2/pull/0.log" Jan 26 17:28:45 crc kubenswrapper[4856]: I0126 17:28:45.934251 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ank5l6_6521dc23-8f4e-452f-ae3e-167424fa3ed2/util/0.log" Jan 26 17:28:45 crc kubenswrapper[4856]: I0126 17:28:45.954036 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ank5l6_6521dc23-8f4e-452f-ae3e-167424fa3ed2/pull/0.log" Jan 26 17:28:45 crc kubenswrapper[4856]: I0126 17:28:45.960684 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931ank5l6_6521dc23-8f4e-452f-ae3e-167424fa3ed2/extract/0.log" Jan 26 17:28:46 crc kubenswrapper[4856]: I0126 17:28:46.097738 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fxrgwg_7105e655-ab8e-4fc0-b205-0bafaa6d7d91/util/0.log" Jan 26 17:28:46 crc kubenswrapper[4856]: I0126 17:28:46.282138 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fxrgwg_7105e655-ab8e-4fc0-b205-0bafaa6d7d91/pull/0.log" Jan 26 17:28:46 crc kubenswrapper[4856]: I0126 17:28:46.282984 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fxrgwg_7105e655-ab8e-4fc0-b205-0bafaa6d7d91/pull/0.log" Jan 26 17:28:46 crc kubenswrapper[4856]: I0126 17:28:46.287475 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fxrgwg_7105e655-ab8e-4fc0-b205-0bafaa6d7d91/util/0.log" Jan 26 17:28:46 crc kubenswrapper[4856]: I0126 17:28:46.407242 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fxrgwg_7105e655-ab8e-4fc0-b205-0bafaa6d7d91/util/0.log" Jan 26 17:28:46 crc kubenswrapper[4856]: I0126 17:28:46.460372 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fxrgwg_7105e655-ab8e-4fc0-b205-0bafaa6d7d91/pull/0.log" Jan 26 17:28:46 crc kubenswrapper[4856]: I0126 17:28:46.464896 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fxrgwg_7105e655-ab8e-4fc0-b205-0bafaa6d7d91/extract/0.log" Jan 26 17:28:46 crc kubenswrapper[4856]: I0126 17:28:46.586286 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5e25mvq_26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a/util/0.log" Jan 26 17:28:46 crc kubenswrapper[4856]: I0126 17:28:46.729152 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5e25mvq_26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a/util/0.log" Jan 26 17:28:46 crc kubenswrapper[4856]: I0126 17:28:46.730800 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5e25mvq_26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a/pull/0.log" Jan 26 17:28:46 crc kubenswrapper[4856]: I0126 17:28:46.763727 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5e25mvq_26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a/pull/0.log" Jan 26 17:28:46 crc kubenswrapper[4856]: I0126 17:28:46.890350 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5e25mvq_26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a/util/0.log" Jan 26 17:28:46 crc kubenswrapper[4856]: I0126 17:28:46.892383 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5e25mvq_26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a/extract/0.log" Jan 26 17:28:46 crc kubenswrapper[4856]: I0126 17:28:46.904684 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5e25mvq_26dfca06-7f67-45e3-ba7c-fb3b0a4ce25a/pull/0.log" Jan 26 17:28:47 crc kubenswrapper[4856]: I0126 17:28:47.031913 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08924r9_64c65d72-3459-4893-a33a-9033e12f188a/util/0.log" Jan 26 17:28:47 crc kubenswrapper[4856]: I0126 17:28:47.211497 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08924r9_64c65d72-3459-4893-a33a-9033e12f188a/util/0.log" Jan 26 17:28:47 crc kubenswrapper[4856]: I0126 17:28:47.212617 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08924r9_64c65d72-3459-4893-a33a-9033e12f188a/pull/0.log" Jan 26 17:28:47 crc kubenswrapper[4856]: I0126 17:28:47.242660 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08924r9_64c65d72-3459-4893-a33a-9033e12f188a/pull/0.log" Jan 26 17:28:47 crc kubenswrapper[4856]: I0126 17:28:47.409089 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08924r9_64c65d72-3459-4893-a33a-9033e12f188a/util/0.log" Jan 26 17:28:47 crc kubenswrapper[4856]: I0126 17:28:47.424215 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08924r9_64c65d72-3459-4893-a33a-9033e12f188a/extract/0.log" Jan 26 17:28:47 crc kubenswrapper[4856]: I0126 17:28:47.439315 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08924r9_64c65d72-3459-4893-a33a-9033e12f188a/pull/0.log" Jan 26 17:28:47 crc kubenswrapper[4856]: I0126 17:28:47.654471 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-hwx7g_a26e6806-50dc-4680-a6d4-01f73fce49da/extract-utilities/0.log" Jan 26 17:28:47 crc kubenswrapper[4856]: I0126 17:28:47.864116 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-hwx7g_a26e6806-50dc-4680-a6d4-01f73fce49da/extract-content/0.log" Jan 26 17:28:47 crc kubenswrapper[4856]: I0126 17:28:47.873037 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-hwx7g_a26e6806-50dc-4680-a6d4-01f73fce49da/extract-utilities/0.log" Jan 26 17:28:47 crc kubenswrapper[4856]: I0126 17:28:47.877360 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-hwx7g_a26e6806-50dc-4680-a6d4-01f73fce49da/extract-content/0.log" Jan 26 17:28:48 crc kubenswrapper[4856]: I0126 17:28:48.052150 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-hwx7g_a26e6806-50dc-4680-a6d4-01f73fce49da/extract-content/0.log" Jan 26 17:28:48 crc kubenswrapper[4856]: I0126 17:28:48.069110 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-hwx7g_a26e6806-50dc-4680-a6d4-01f73fce49da/extract-utilities/0.log" Jan 26 17:28:48 crc kubenswrapper[4856]: I0126 17:28:48.175269 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-hwx7g_a26e6806-50dc-4680-a6d4-01f73fce49da/registry-server/0.log" Jan 26 17:28:48 crc kubenswrapper[4856]: I0126 17:28:48.282403 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gdp2n_4327b726-2edc-40ad-ac96-b19a7e020048/extract-utilities/0.log" Jan 26 17:28:48 crc kubenswrapper[4856]: I0126 17:28:48.440755 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gdp2n_4327b726-2edc-40ad-ac96-b19a7e020048/extract-utilities/0.log" Jan 26 17:28:48 crc kubenswrapper[4856]: I0126 17:28:48.441753 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gdp2n_4327b726-2edc-40ad-ac96-b19a7e020048/extract-content/0.log" Jan 26 17:28:48 crc kubenswrapper[4856]: I0126 17:28:48.443663 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gdp2n_4327b726-2edc-40ad-ac96-b19a7e020048/extract-content/0.log" Jan 26 17:28:48 crc kubenswrapper[4856]: I0126 17:28:48.569187 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gdp2n_4327b726-2edc-40ad-ac96-b19a7e020048/extract-utilities/0.log" Jan 26 17:28:48 crc kubenswrapper[4856]: I0126 17:28:48.627880 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gdp2n_4327b726-2edc-40ad-ac96-b19a7e020048/extract-content/0.log" Jan 26 17:28:48 crc kubenswrapper[4856]: I0126 17:28:48.864863 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-tdtfh_566ca894-037a-4b73-95d4-a6246c7c851a/marketplace-operator/0.log" Jan 26 17:28:48 crc kubenswrapper[4856]: I0126 17:28:48.880265 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lfhpz_8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee/extract-utilities/0.log" Jan 26 17:28:49 crc kubenswrapper[4856]: I0126 17:28:49.131191 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-gdp2n_4327b726-2edc-40ad-ac96-b19a7e020048/registry-server/0.log" Jan 26 17:28:49 crc kubenswrapper[4856]: I0126 17:28:49.169466 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lfhpz_8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee/extract-utilities/0.log" Jan 26 17:28:49 crc kubenswrapper[4856]: I0126 17:28:49.227732 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lfhpz_8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee/extract-content/0.log" Jan 26 17:28:49 crc kubenswrapper[4856]: I0126 17:28:49.257297 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lfhpz_8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee/extract-content/0.log" Jan 26 17:28:49 crc kubenswrapper[4856]: I0126 17:28:49.357469 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lfhpz_8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee/extract-content/0.log" Jan 26 17:28:49 crc kubenswrapper[4856]: I0126 17:28:49.394907 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lfhpz_8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee/extract-utilities/0.log" Jan 26 17:28:49 crc kubenswrapper[4856]: I0126 17:28:49.868941 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-lfhpz_8cd36133-7a25-4dae-83a4-bbd0fbf1f2ee/registry-server/0.log" Jan 26 17:28:57 crc kubenswrapper[4856]: I0126 17:28:57.395284 4856 scope.go:117] "RemoveContainer" containerID="b8175a0e79754a858867d9a98f2aa6c52214536db6005b6724cf907eb7a891ee" Jan 26 17:28:57 crc kubenswrapper[4856]: E0126 17:28:57.396275 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xm9cq_openshift-machine-config-operator(63c75ede-5170-4db0-811b-5217ef8d72b3)\"" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" Jan 26 17:29:01 crc kubenswrapper[4856]: I0126 17:29:01.849557 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-cq2gx_e31d2d53-8992-45e3-98aa-24ea73236248/prometheus-operator/0.log" Jan 26 17:29:01 crc kubenswrapper[4856]: I0126 17:29:01.873773 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-68d9fdc4dd-cf7wn_7c88687f-1304-4709-b148-a196f0d0190d/prometheus-operator-admission-webhook/0.log" Jan 26 17:29:01 crc kubenswrapper[4856]: I0126 17:29:01.893942 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-68d9fdc4dd-jbq25_766f50ba-0751-4f25-a6db-3b7195e72f55/prometheus-operator-admission-webhook/0.log" Jan 26 17:29:02 crc kubenswrapper[4856]: I0126 17:29:02.037720 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-5bmfp_bd7597f2-d44b-4e1b-ac60-b409985e3351/perses-operator/0.log" Jan 26 17:29:02 crc kubenswrapper[4856]: I0126 17:29:02.052984 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-fpn2h_a4ae7646-2afb-4ada-b8a4-d20a69f87949/operator/0.log" Jan 26 17:29:11 crc kubenswrapper[4856]: I0126 17:29:11.395085 4856 scope.go:117] "RemoveContainer" containerID="b8175a0e79754a858867d9a98f2aa6c52214536db6005b6724cf907eb7a891ee" Jan 26 17:29:11 crc kubenswrapper[4856]: E0126 17:29:11.396083 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xm9cq_openshift-machine-config-operator(63c75ede-5170-4db0-811b-5217ef8d72b3)\"" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" Jan 26 17:29:23 crc kubenswrapper[4856]: I0126 17:29:23.396200 4856 scope.go:117] "RemoveContainer" containerID="b8175a0e79754a858867d9a98f2aa6c52214536db6005b6724cf907eb7a891ee" Jan 26 17:29:23 crc kubenswrapper[4856]: E0126 17:29:23.397588 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xm9cq_openshift-machine-config-operator(63c75ede-5170-4db0-811b-5217ef8d72b3)\"" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" Jan 26 17:29:38 crc kubenswrapper[4856]: I0126 17:29:38.396196 4856 scope.go:117] "RemoveContainer" containerID="b8175a0e79754a858867d9a98f2aa6c52214536db6005b6724cf907eb7a891ee" Jan 26 17:29:38 crc kubenswrapper[4856]: E0126 17:29:38.397085 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xm9cq_openshift-machine-config-operator(63c75ede-5170-4db0-811b-5217ef8d72b3)\"" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" Jan 26 17:29:53 crc kubenswrapper[4856]: I0126 17:29:53.395739 4856 scope.go:117] "RemoveContainer" containerID="b8175a0e79754a858867d9a98f2aa6c52214536db6005b6724cf907eb7a891ee" Jan 26 17:29:53 crc kubenswrapper[4856]: E0126 17:29:53.398168 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xm9cq_openshift-machine-config-operator(63c75ede-5170-4db0-811b-5217ef8d72b3)\"" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" Jan 26 17:30:00 crc kubenswrapper[4856]: I0126 17:30:00.159396 4856 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490810-lg76m"] Jan 26 17:30:00 crc kubenswrapper[4856]: I0126 17:30:00.161093 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490810-lg76m" Jan 26 17:30:00 crc kubenswrapper[4856]: I0126 17:30:00.164537 4856 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 26 17:30:00 crc kubenswrapper[4856]: I0126 17:30:00.164522 4856 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 26 17:30:00 crc kubenswrapper[4856]: I0126 17:30:00.185908 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490810-lg76m"] Jan 26 17:30:00 crc kubenswrapper[4856]: I0126 17:30:00.262647 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnmfh\" (UniqueName: \"kubernetes.io/projected/886317f6-fcda-4393-a343-e9dda96dbc1f-kube-api-access-qnmfh\") pod \"collect-profiles-29490810-lg76m\" (UID: \"886317f6-fcda-4393-a343-e9dda96dbc1f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490810-lg76m" Jan 26 17:30:00 crc kubenswrapper[4856]: I0126 17:30:00.263027 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/886317f6-fcda-4393-a343-e9dda96dbc1f-config-volume\") pod \"collect-profiles-29490810-lg76m\" (UID: \"886317f6-fcda-4393-a343-e9dda96dbc1f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490810-lg76m" Jan 26 17:30:00 crc kubenswrapper[4856]: I0126 17:30:00.263062 4856 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/886317f6-fcda-4393-a343-e9dda96dbc1f-secret-volume\") pod \"collect-profiles-29490810-lg76m\" (UID: \"886317f6-fcda-4393-a343-e9dda96dbc1f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490810-lg76m" Jan 26 17:30:00 crc kubenswrapper[4856]: I0126 17:30:00.365092 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnmfh\" (UniqueName: \"kubernetes.io/projected/886317f6-fcda-4393-a343-e9dda96dbc1f-kube-api-access-qnmfh\") pod \"collect-profiles-29490810-lg76m\" (UID: \"886317f6-fcda-4393-a343-e9dda96dbc1f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490810-lg76m" Jan 26 17:30:00 crc kubenswrapper[4856]: I0126 17:30:00.365670 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/886317f6-fcda-4393-a343-e9dda96dbc1f-config-volume\") pod \"collect-profiles-29490810-lg76m\" (UID: \"886317f6-fcda-4393-a343-e9dda96dbc1f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490810-lg76m" Jan 26 17:30:00 crc kubenswrapper[4856]: I0126 17:30:00.365793 4856 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/886317f6-fcda-4393-a343-e9dda96dbc1f-secret-volume\") pod \"collect-profiles-29490810-lg76m\" (UID: \"886317f6-fcda-4393-a343-e9dda96dbc1f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490810-lg76m" Jan 26 17:30:00 crc kubenswrapper[4856]: I0126 17:30:00.367211 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/886317f6-fcda-4393-a343-e9dda96dbc1f-config-volume\") pod \"collect-profiles-29490810-lg76m\" (UID: \"886317f6-fcda-4393-a343-e9dda96dbc1f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490810-lg76m" Jan 26 17:30:00 crc kubenswrapper[4856]: I0126 17:30:00.374838 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/886317f6-fcda-4393-a343-e9dda96dbc1f-secret-volume\") pod \"collect-profiles-29490810-lg76m\" (UID: \"886317f6-fcda-4393-a343-e9dda96dbc1f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490810-lg76m" Jan 26 17:30:00 crc kubenswrapper[4856]: I0126 17:30:00.386027 4856 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnmfh\" (UniqueName: \"kubernetes.io/projected/886317f6-fcda-4393-a343-e9dda96dbc1f-kube-api-access-qnmfh\") pod \"collect-profiles-29490810-lg76m\" (UID: \"886317f6-fcda-4393-a343-e9dda96dbc1f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490810-lg76m" Jan 26 17:30:00 crc kubenswrapper[4856]: I0126 17:30:00.488644 4856 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490810-lg76m" Jan 26 17:30:00 crc kubenswrapper[4856]: I0126 17:30:00.529108 4856 generic.go:334] "Generic (PLEG): container finished" podID="0190f6f3-2762-486f-8648-e59c9e5179ad" containerID="1c6e1c2b24da05f9138f08890b3f71dfa17ade72ec99276a5860cbc9c8b26e99" exitCode=0 Jan 26 17:30:00 crc kubenswrapper[4856]: I0126 17:30:00.529232 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pxr8g/must-gather-xhb5b" event={"ID":"0190f6f3-2762-486f-8648-e59c9e5179ad","Type":"ContainerDied","Data":"1c6e1c2b24da05f9138f08890b3f71dfa17ade72ec99276a5860cbc9c8b26e99"} Jan 26 17:30:00 crc kubenswrapper[4856]: I0126 17:30:00.530145 4856 scope.go:117] "RemoveContainer" containerID="1c6e1c2b24da05f9138f08890b3f71dfa17ade72ec99276a5860cbc9c8b26e99" Jan 26 17:30:01 crc kubenswrapper[4856]: I0126 17:30:01.000687 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-pxr8g_must-gather-xhb5b_0190f6f3-2762-486f-8648-e59c9e5179ad/gather/0.log" Jan 26 17:30:01 crc kubenswrapper[4856]: I0126 17:30:01.016031 4856 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490810-lg76m"] Jan 26 17:30:01 crc kubenswrapper[4856]: I0126 17:30:01.539383 4856 generic.go:334] "Generic (PLEG): container finished" podID="886317f6-fcda-4393-a343-e9dda96dbc1f" containerID="6d9a89251f6438ddca07d6a74a8d062369c94886db42eccc2d70e510259e9cf6" exitCode=0 Jan 26 17:30:01 crc kubenswrapper[4856]: I0126 17:30:01.539439 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490810-lg76m" event={"ID":"886317f6-fcda-4393-a343-e9dda96dbc1f","Type":"ContainerDied","Data":"6d9a89251f6438ddca07d6a74a8d062369c94886db42eccc2d70e510259e9cf6"} Jan 26 17:30:01 crc kubenswrapper[4856]: I0126 17:30:01.539501 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490810-lg76m" event={"ID":"886317f6-fcda-4393-a343-e9dda96dbc1f","Type":"ContainerStarted","Data":"eca76d55256c0050d6dac39b2a19159f5841b8a78d5df6bb7d1baf089d8d380e"} Jan 26 17:30:02 crc kubenswrapper[4856]: I0126 17:30:02.850171 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490810-lg76m" Jan 26 17:30:03 crc kubenswrapper[4856]: I0126 17:30:03.000242 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/886317f6-fcda-4393-a343-e9dda96dbc1f-secret-volume\") pod \"886317f6-fcda-4393-a343-e9dda96dbc1f\" (UID: \"886317f6-fcda-4393-a343-e9dda96dbc1f\") " Jan 26 17:30:03 crc kubenswrapper[4856]: I0126 17:30:03.000328 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/886317f6-fcda-4393-a343-e9dda96dbc1f-config-volume\") pod \"886317f6-fcda-4393-a343-e9dda96dbc1f\" (UID: \"886317f6-fcda-4393-a343-e9dda96dbc1f\") " Jan 26 17:30:03 crc kubenswrapper[4856]: I0126 17:30:03.000452 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnmfh\" (UniqueName: \"kubernetes.io/projected/886317f6-fcda-4393-a343-e9dda96dbc1f-kube-api-access-qnmfh\") pod \"886317f6-fcda-4393-a343-e9dda96dbc1f\" (UID: \"886317f6-fcda-4393-a343-e9dda96dbc1f\") " Jan 26 17:30:03 crc kubenswrapper[4856]: I0126 17:30:03.001083 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/886317f6-fcda-4393-a343-e9dda96dbc1f-config-volume" (OuterVolumeSpecName: "config-volume") pod "886317f6-fcda-4393-a343-e9dda96dbc1f" (UID: "886317f6-fcda-4393-a343-e9dda96dbc1f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 17:30:03 crc kubenswrapper[4856]: I0126 17:30:03.005868 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/886317f6-fcda-4393-a343-e9dda96dbc1f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "886317f6-fcda-4393-a343-e9dda96dbc1f" (UID: "886317f6-fcda-4393-a343-e9dda96dbc1f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 17:30:03 crc kubenswrapper[4856]: I0126 17:30:03.006014 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/886317f6-fcda-4393-a343-e9dda96dbc1f-kube-api-access-qnmfh" (OuterVolumeSpecName: "kube-api-access-qnmfh") pod "886317f6-fcda-4393-a343-e9dda96dbc1f" (UID: "886317f6-fcda-4393-a343-e9dda96dbc1f"). InnerVolumeSpecName "kube-api-access-qnmfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:30:03 crc kubenswrapper[4856]: I0126 17:30:03.102021 4856 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/886317f6-fcda-4393-a343-e9dda96dbc1f-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 26 17:30:03 crc kubenswrapper[4856]: I0126 17:30:03.102071 4856 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/886317f6-fcda-4393-a343-e9dda96dbc1f-config-volume\") on node \"crc\" DevicePath \"\"" Jan 26 17:30:03 crc kubenswrapper[4856]: I0126 17:30:03.102090 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnmfh\" (UniqueName: \"kubernetes.io/projected/886317f6-fcda-4393-a343-e9dda96dbc1f-kube-api-access-qnmfh\") on node \"crc\" DevicePath \"\"" Jan 26 17:30:03 crc kubenswrapper[4856]: I0126 17:30:03.559758 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490810-lg76m" event={"ID":"886317f6-fcda-4393-a343-e9dda96dbc1f","Type":"ContainerDied","Data":"eca76d55256c0050d6dac39b2a19159f5841b8a78d5df6bb7d1baf089d8d380e"} Jan 26 17:30:03 crc kubenswrapper[4856]: I0126 17:30:03.559818 4856 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eca76d55256c0050d6dac39b2a19159f5841b8a78d5df6bb7d1baf089d8d380e" Jan 26 17:30:03 crc kubenswrapper[4856]: I0126 17:30:03.559847 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490810-lg76m" Jan 26 17:30:07 crc kubenswrapper[4856]: I0126 17:30:07.396373 4856 scope.go:117] "RemoveContainer" containerID="b8175a0e79754a858867d9a98f2aa6c52214536db6005b6724cf907eb7a891ee" Jan 26 17:30:07 crc kubenswrapper[4856]: E0126 17:30:07.397275 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xm9cq_openshift-machine-config-operator(63c75ede-5170-4db0-811b-5217ef8d72b3)\"" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" Jan 26 17:30:07 crc kubenswrapper[4856]: I0126 17:30:07.919701 4856 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-pxr8g/must-gather-xhb5b"] Jan 26 17:30:07 crc kubenswrapper[4856]: I0126 17:30:07.920380 4856 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-pxr8g/must-gather-xhb5b" podUID="0190f6f3-2762-486f-8648-e59c9e5179ad" containerName="copy" containerID="cri-o://3f464ae4156c51b0e76930516abce0275686a5ccb1e2b35e0204d8a9e05e513a" gracePeriod=2 Jan 26 17:30:07 crc kubenswrapper[4856]: I0126 17:30:07.925822 4856 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-pxr8g/must-gather-xhb5b"] Jan 26 17:30:08 crc kubenswrapper[4856]: I0126 17:30:08.236376 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-pxr8g_must-gather-xhb5b_0190f6f3-2762-486f-8648-e59c9e5179ad/copy/0.log" Jan 26 17:30:08 crc kubenswrapper[4856]: I0126 17:30:08.237169 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pxr8g/must-gather-xhb5b" Jan 26 17:30:08 crc kubenswrapper[4856]: I0126 17:30:08.382574 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0190f6f3-2762-486f-8648-e59c9e5179ad-must-gather-output\") pod \"0190f6f3-2762-486f-8648-e59c9e5179ad\" (UID: \"0190f6f3-2762-486f-8648-e59c9e5179ad\") " Jan 26 17:30:08 crc kubenswrapper[4856]: I0126 17:30:08.382640 4856 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vms6d\" (UniqueName: \"kubernetes.io/projected/0190f6f3-2762-486f-8648-e59c9e5179ad-kube-api-access-vms6d\") pod \"0190f6f3-2762-486f-8648-e59c9e5179ad\" (UID: \"0190f6f3-2762-486f-8648-e59c9e5179ad\") " Jan 26 17:30:08 crc kubenswrapper[4856]: I0126 17:30:08.390447 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0190f6f3-2762-486f-8648-e59c9e5179ad-kube-api-access-vms6d" (OuterVolumeSpecName: "kube-api-access-vms6d") pod "0190f6f3-2762-486f-8648-e59c9e5179ad" (UID: "0190f6f3-2762-486f-8648-e59c9e5179ad"). InnerVolumeSpecName "kube-api-access-vms6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 17:30:08 crc kubenswrapper[4856]: I0126 17:30:08.442354 4856 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0190f6f3-2762-486f-8648-e59c9e5179ad-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "0190f6f3-2762-486f-8648-e59c9e5179ad" (UID: "0190f6f3-2762-486f-8648-e59c9e5179ad"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 17:30:08 crc kubenswrapper[4856]: I0126 17:30:08.484707 4856 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/0190f6f3-2762-486f-8648-e59c9e5179ad-must-gather-output\") on node \"crc\" DevicePath \"\"" Jan 26 17:30:08 crc kubenswrapper[4856]: I0126 17:30:08.484781 4856 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vms6d\" (UniqueName: \"kubernetes.io/projected/0190f6f3-2762-486f-8648-e59c9e5179ad-kube-api-access-vms6d\") on node \"crc\" DevicePath \"\"" Jan 26 17:30:08 crc kubenswrapper[4856]: I0126 17:30:08.607679 4856 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-pxr8g_must-gather-xhb5b_0190f6f3-2762-486f-8648-e59c9e5179ad/copy/0.log" Jan 26 17:30:08 crc kubenswrapper[4856]: I0126 17:30:08.608463 4856 generic.go:334] "Generic (PLEG): container finished" podID="0190f6f3-2762-486f-8648-e59c9e5179ad" containerID="3f464ae4156c51b0e76930516abce0275686a5ccb1e2b35e0204d8a9e05e513a" exitCode=143 Jan 26 17:30:08 crc kubenswrapper[4856]: I0126 17:30:08.608566 4856 scope.go:117] "RemoveContainer" containerID="3f464ae4156c51b0e76930516abce0275686a5ccb1e2b35e0204d8a9e05e513a" Jan 26 17:30:08 crc kubenswrapper[4856]: I0126 17:30:08.608629 4856 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pxr8g/must-gather-xhb5b" Jan 26 17:30:08 crc kubenswrapper[4856]: I0126 17:30:08.639383 4856 scope.go:117] "RemoveContainer" containerID="1c6e1c2b24da05f9138f08890b3f71dfa17ade72ec99276a5860cbc9c8b26e99" Jan 26 17:30:08 crc kubenswrapper[4856]: I0126 17:30:08.690816 4856 scope.go:117] "RemoveContainer" containerID="3f464ae4156c51b0e76930516abce0275686a5ccb1e2b35e0204d8a9e05e513a" Jan 26 17:30:08 crc kubenswrapper[4856]: E0126 17:30:08.691180 4856 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f464ae4156c51b0e76930516abce0275686a5ccb1e2b35e0204d8a9e05e513a\": container with ID starting with 3f464ae4156c51b0e76930516abce0275686a5ccb1e2b35e0204d8a9e05e513a not found: ID does not exist" containerID="3f464ae4156c51b0e76930516abce0275686a5ccb1e2b35e0204d8a9e05e513a" Jan 26 17:30:08 crc kubenswrapper[4856]: I0126 17:30:08.691224 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f464ae4156c51b0e76930516abce0275686a5ccb1e2b35e0204d8a9e05e513a"} err="failed to get container status \"3f464ae4156c51b0e76930516abce0275686a5ccb1e2b35e0204d8a9e05e513a\": rpc error: code = NotFound desc = could not find container \"3f464ae4156c51b0e76930516abce0275686a5ccb1e2b35e0204d8a9e05e513a\": container with ID starting with 3f464ae4156c51b0e76930516abce0275686a5ccb1e2b35e0204d8a9e05e513a not found: ID does not exist" Jan 26 17:30:08 crc kubenswrapper[4856]: I0126 17:30:08.691245 4856 scope.go:117] "RemoveContainer" containerID="1c6e1c2b24da05f9138f08890b3f71dfa17ade72ec99276a5860cbc9c8b26e99" Jan 26 17:30:08 crc kubenswrapper[4856]: E0126 17:30:08.691454 4856 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c6e1c2b24da05f9138f08890b3f71dfa17ade72ec99276a5860cbc9c8b26e99\": container with ID starting with 1c6e1c2b24da05f9138f08890b3f71dfa17ade72ec99276a5860cbc9c8b26e99 not found: ID does not exist" containerID="1c6e1c2b24da05f9138f08890b3f71dfa17ade72ec99276a5860cbc9c8b26e99" Jan 26 17:30:08 crc kubenswrapper[4856]: I0126 17:30:08.691500 4856 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c6e1c2b24da05f9138f08890b3f71dfa17ade72ec99276a5860cbc9c8b26e99"} err="failed to get container status \"1c6e1c2b24da05f9138f08890b3f71dfa17ade72ec99276a5860cbc9c8b26e99\": rpc error: code = NotFound desc = could not find container \"1c6e1c2b24da05f9138f08890b3f71dfa17ade72ec99276a5860cbc9c8b26e99\": container with ID starting with 1c6e1c2b24da05f9138f08890b3f71dfa17ade72ec99276a5860cbc9c8b26e99 not found: ID does not exist" Jan 26 17:30:09 crc kubenswrapper[4856]: I0126 17:30:09.439316 4856 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0190f6f3-2762-486f-8648-e59c9e5179ad" path="/var/lib/kubelet/pods/0190f6f3-2762-486f-8648-e59c9e5179ad/volumes" Jan 26 17:30:19 crc kubenswrapper[4856]: I0126 17:30:19.401361 4856 scope.go:117] "RemoveContainer" containerID="b8175a0e79754a858867d9a98f2aa6c52214536db6005b6724cf907eb7a891ee" Jan 26 17:30:19 crc kubenswrapper[4856]: E0126 17:30:19.404216 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xm9cq_openshift-machine-config-operator(63c75ede-5170-4db0-811b-5217ef8d72b3)\"" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" Jan 26 17:30:31 crc kubenswrapper[4856]: I0126 17:30:31.395069 4856 scope.go:117] "RemoveContainer" containerID="b8175a0e79754a858867d9a98f2aa6c52214536db6005b6724cf907eb7a891ee" Jan 26 17:30:31 crc kubenswrapper[4856]: E0126 17:30:31.396040 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xm9cq_openshift-machine-config-operator(63c75ede-5170-4db0-811b-5217ef8d72b3)\"" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" Jan 26 17:30:43 crc kubenswrapper[4856]: I0126 17:30:43.983692 4856 scope.go:117] "RemoveContainer" containerID="bc8203a7d47ee43722cf1047b5eda9c60a27c27e6ee323351ad4ef557fd2f359" Jan 26 17:30:44 crc kubenswrapper[4856]: I0126 17:30:44.002826 4856 scope.go:117] "RemoveContainer" containerID="ea789d382e8ee7680a387d4c0ac031fe25c48fa1835d8928284e77f51f936f6a" Jan 26 17:30:44 crc kubenswrapper[4856]: I0126 17:30:44.021770 4856 scope.go:117] "RemoveContainer" containerID="26c1821c4f23578ac6a0c1415d6730caff4fddaf6502ee746d86ad133657cf42" Jan 26 17:30:46 crc kubenswrapper[4856]: I0126 17:30:46.395414 4856 scope.go:117] "RemoveContainer" containerID="b8175a0e79754a858867d9a98f2aa6c52214536db6005b6724cf907eb7a891ee" Jan 26 17:30:46 crc kubenswrapper[4856]: E0126 17:30:46.396018 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xm9cq_openshift-machine-config-operator(63c75ede-5170-4db0-811b-5217ef8d72b3)\"" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" Jan 26 17:30:57 crc kubenswrapper[4856]: I0126 17:30:57.395720 4856 scope.go:117] "RemoveContainer" containerID="b8175a0e79754a858867d9a98f2aa6c52214536db6005b6724cf907eb7a891ee" Jan 26 17:30:57 crc kubenswrapper[4856]: E0126 17:30:57.396460 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xm9cq_openshift-machine-config-operator(63c75ede-5170-4db0-811b-5217ef8d72b3)\"" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" Jan 26 17:31:10 crc kubenswrapper[4856]: I0126 17:31:10.395075 4856 scope.go:117] "RemoveContainer" containerID="b8175a0e79754a858867d9a98f2aa6c52214536db6005b6724cf907eb7a891ee" Jan 26 17:31:10 crc kubenswrapper[4856]: E0126 17:31:10.396057 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xm9cq_openshift-machine-config-operator(63c75ede-5170-4db0-811b-5217ef8d72b3)\"" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" Jan 26 17:31:23 crc kubenswrapper[4856]: I0126 17:31:23.395046 4856 scope.go:117] "RemoveContainer" containerID="b8175a0e79754a858867d9a98f2aa6c52214536db6005b6724cf907eb7a891ee" Jan 26 17:31:23 crc kubenswrapper[4856]: E0126 17:31:23.396001 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xm9cq_openshift-machine-config-operator(63c75ede-5170-4db0-811b-5217ef8d72b3)\"" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" Jan 26 17:31:35 crc kubenswrapper[4856]: I0126 17:31:35.410882 4856 scope.go:117] "RemoveContainer" containerID="b8175a0e79754a858867d9a98f2aa6c52214536db6005b6724cf907eb7a891ee" Jan 26 17:31:35 crc kubenswrapper[4856]: E0126 17:31:35.411598 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xm9cq_openshift-machine-config-operator(63c75ede-5170-4db0-811b-5217ef8d72b3)\"" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" Jan 26 17:31:50 crc kubenswrapper[4856]: I0126 17:31:50.395356 4856 scope.go:117] "RemoveContainer" containerID="b8175a0e79754a858867d9a98f2aa6c52214536db6005b6724cf907eb7a891ee" Jan 26 17:31:50 crc kubenswrapper[4856]: E0126 17:31:50.396252 4856 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xm9cq_openshift-machine-config-operator(63c75ede-5170-4db0-811b-5217ef8d72b3)\"" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" podUID="63c75ede-5170-4db0-811b-5217ef8d72b3" Jan 26 17:32:04 crc kubenswrapper[4856]: I0126 17:32:04.395359 4856 scope.go:117] "RemoveContainer" containerID="b8175a0e79754a858867d9a98f2aa6c52214536db6005b6724cf907eb7a891ee" Jan 26 17:32:05 crc kubenswrapper[4856]: I0126 17:32:05.602909 4856 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xm9cq" event={"ID":"63c75ede-5170-4db0-811b-5217ef8d72b3","Type":"ContainerStarted","Data":"d8ad8f8ca78c63277c4977095b5c49dedf5824f809bcd40f7a2a1038751c894c"} var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515135722527024456 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015135722527017373 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015135716075016517 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015135716075015467 5ustar corecore